docs: Fix broken links in AI docs (#30872)

This PR fixes some broken links in the AI docs.

Release Notes:

- N/A
This commit is contained in:
Marshall Bowers 2025-05-17 13:54:42 +02:00 committed by GitHub
parent e518941445
commit 175ce05fd1
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
2 changed files with 3 additions and 3 deletions

View file

@ -5,7 +5,7 @@ You can use it for various tasks, such as generating code, asking questions abou
To open the Agent Panel, use the `agent: new thread` action in [the Command Palette](../getting-started.md#command-palette) or click the ✨ (sparkles) icon in the status bar.
If you're using the Agent Panel for the first time, you'll need to [configure at least one LLM provider](./ai/configuration.md).
If you're using the Agent Panel for the first time, you'll need to [configure at least one LLM provider](./configuration.md).
## Overview {#overview}
@ -78,7 +78,7 @@ You can also do this at any time with an ongoing thread via the "Agent Options"
## Changing Models {#changing-models}
After you've configured your LLM providers—either via [a custom API key](./custom-api-keys.md) or through [Zed's hosted models](./models.md)—you can switch between them by clicking on the model selector on the message editor or by using the {#kb agent::ToggleModelSelector} keybinding.
After you've configured your LLM providers—either via [a custom API key](./configuration.md#use-your-own-keys) or through [Zed's hosted models](./models.md)—you can switch between them by clicking on the model selector on the message editor or by using the {#kb agent::ToggleModelSelector} keybinding.
## Using Tools {#using-tools}

View file

@ -391,7 +391,7 @@ Example configuration:
You can configure additional models that will be used to perform inline assists in parallel.
When you do this, the inline assist UI will surface controls to cycle between the alternatives generated by each model.
The models you specify here are always used in _addition_ to your [default model](./ai/configuration.md#default-model).
The models you specify here are always used in _addition_ to your [default model](#default-model).
For example, the following configuration will generate two outputs for every assist.
One with Claude 3.7 Sonnet, and one with GPT-4o.