mod agent_profile; use std::sync::Arc; use ::open_ai::Model as OpenAiModel; use anthropic::Model as AnthropicModel; use anyhow::{Result, bail}; use deepseek::Model as DeepseekModel; use feature_flags::{AgentStreamEditsFeatureFlag, Assistant2FeatureFlag, FeatureFlagAppExt}; use gpui::{App, Pixels}; use indexmap::IndexMap; use language_model::{CloudModel, LanguageModel}; use lmstudio::Model as LmStudioModel; use ollama::Model as OllamaModel; use schemars::{JsonSchema, schema::Schema}; use serde::{Deserialize, Serialize}; use settings::{Settings, SettingsSources}; pub use crate::agent_profile::*; #[derive(Copy, Clone, Default, Debug, Serialize, Deserialize, JsonSchema)] #[serde(rename_all = "snake_case")] pub enum AssistantDockPosition { Left, #[default] Right, Bottom, } #[derive(Copy, Clone, Default, Debug, Serialize, Deserialize, JsonSchema, PartialEq, Eq)] #[serde(rename_all = "snake_case")] pub enum NotifyWhenAgentWaiting { #[default] PrimaryScreen, AllScreens, Never, } #[derive(Clone, Debug, Serialize, Deserialize, JsonSchema, PartialEq)] #[serde(tag = "name", rename_all = "snake_case")] pub enum AssistantProviderContentV1 { #[serde(rename = "zed.dev")] ZedDotDev { default_model: Option }, #[serde(rename = "openai")] OpenAi { default_model: Option, api_url: Option, available_models: Option>, }, #[serde(rename = "anthropic")] Anthropic { default_model: Option, api_url: Option, }, #[serde(rename = "ollama")] Ollama { default_model: Option, api_url: Option, }, #[serde(rename = "lmstudio")] LmStudio { default_model: Option, api_url: Option, }, #[serde(rename = "deepseek")] DeepSeek { default_model: Option, api_url: Option, }, } #[derive(Default, Clone, Debug)] pub struct AssistantSettings { pub enabled: bool, pub button: bool, pub dock: AssistantDockPosition, pub default_width: Pixels, pub default_height: Pixels, pub default_model: LanguageModelSelection, pub inline_assistant_model: Option, pub commit_message_model: Option, pub thread_summary_model: Option, pub inline_alternatives: Vec, pub using_outdated_settings_version: bool, pub enable_experimental_live_diffs: bool, pub default_profile: AgentProfileId, pub profiles: IndexMap, pub always_allow_tool_actions: bool, pub notify_when_agent_waiting: NotifyWhenAgentWaiting, pub stream_edits: bool, pub single_file_review: bool, } impl AssistantSettings { pub fn stream_edits(&self, cx: &App) -> bool { cx.has_flag::() || self.stream_edits } pub fn are_live_diffs_enabled(&self, cx: &App) -> bool { if cx.has_flag::() { return false; } cx.is_staff() || self.enable_experimental_live_diffs } pub fn set_inline_assistant_model(&mut self, provider: String, model: String) { self.inline_assistant_model = Some(LanguageModelSelection { provider, model }); } pub fn set_commit_message_model(&mut self, provider: String, model: String) { self.commit_message_model = Some(LanguageModelSelection { provider, model }); } pub fn set_thread_summary_model(&mut self, provider: String, model: String) { self.thread_summary_model = Some(LanguageModelSelection { provider, model }); } } /// Assistant panel settings #[derive(Clone, Serialize, Deserialize, Debug, Default)] pub struct AssistantSettingsContent { #[serde(flatten)] pub inner: Option, } #[derive(Clone, Serialize, Deserialize, Debug)] #[serde(untagged)] pub enum AssistantSettingsContentInner { Versioned(Box), Legacy(LegacyAssistantSettingsContent), } impl AssistantSettingsContentInner { fn for_v2(content: AssistantSettingsContentV2) -> Self { AssistantSettingsContentInner::Versioned(Box::new(VersionedAssistantSettingsContent::V2( content, ))) } } impl JsonSchema for AssistantSettingsContent { fn schema_name() -> String { VersionedAssistantSettingsContent::schema_name() } fn json_schema(r#gen: &mut schemars::r#gen::SchemaGenerator) -> Schema { VersionedAssistantSettingsContent::json_schema(r#gen) } fn is_referenceable() -> bool { VersionedAssistantSettingsContent::is_referenceable() } } impl AssistantSettingsContent { pub fn is_version_outdated(&self) -> bool { match &self.inner { Some(AssistantSettingsContentInner::Versioned(settings)) => match **settings { VersionedAssistantSettingsContent::V1(_) => true, VersionedAssistantSettingsContent::V2(_) => false, }, Some(AssistantSettingsContentInner::Legacy(_)) => true, None => false, } } fn upgrade(&self) -> AssistantSettingsContentV2 { match &self.inner { Some(AssistantSettingsContentInner::Versioned(settings)) => match **settings { VersionedAssistantSettingsContent::V1(ref settings) => AssistantSettingsContentV2 { enabled: settings.enabled, button: settings.button, dock: settings.dock, default_width: settings.default_width, default_height: settings.default_width, default_model: settings .provider .clone() .and_then(|provider| match provider { AssistantProviderContentV1::ZedDotDev { default_model } => { default_model.map(|model| LanguageModelSelection { provider: "zed.dev".to_string(), model: model.id().to_string(), }) } AssistantProviderContentV1::OpenAi { default_model, .. } => { default_model.map(|model| LanguageModelSelection { provider: "openai".to_string(), model: model.id().to_string(), }) } AssistantProviderContentV1::Anthropic { default_model, .. } => { default_model.map(|model| LanguageModelSelection { provider: "anthropic".to_string(), model: model.id().to_string(), }) } AssistantProviderContentV1::Ollama { default_model, .. } => { default_model.map(|model| LanguageModelSelection { provider: "ollama".to_string(), model: model.id().to_string(), }) } AssistantProviderContentV1::LmStudio { default_model, .. } => { default_model.map(|model| LanguageModelSelection { provider: "lmstudio".to_string(), model: model.id().to_string(), }) } AssistantProviderContentV1::DeepSeek { default_model, .. } => { default_model.map(|model| LanguageModelSelection { provider: "deepseek".to_string(), model: model.id().to_string(), }) } }), inline_assistant_model: None, commit_message_model: None, thread_summary_model: None, inline_alternatives: None, enable_experimental_live_diffs: None, default_profile: None, profiles: None, always_allow_tool_actions: None, notify_when_agent_waiting: None, stream_edits: None, single_file_review: None, }, VersionedAssistantSettingsContent::V2(ref settings) => settings.clone(), }, Some(AssistantSettingsContentInner::Legacy(settings)) => AssistantSettingsContentV2 { enabled: None, button: settings.button, dock: settings.dock, default_width: settings.default_width, default_height: settings.default_height, default_model: Some(LanguageModelSelection { provider: "openai".to_string(), model: settings .default_open_ai_model .clone() .unwrap_or_default() .id() .to_string(), }), inline_assistant_model: None, commit_message_model: None, thread_summary_model: None, inline_alternatives: None, enable_experimental_live_diffs: None, default_profile: None, profiles: None, always_allow_tool_actions: None, notify_when_agent_waiting: None, stream_edits: None, single_file_review: None, }, None => AssistantSettingsContentV2::default(), } } pub fn set_dock(&mut self, dock: AssistantDockPosition) { match &mut self.inner { Some(AssistantSettingsContentInner::Versioned(settings)) => match **settings { VersionedAssistantSettingsContent::V1(ref mut settings) => { settings.dock = Some(dock); } VersionedAssistantSettingsContent::V2(ref mut settings) => { settings.dock = Some(dock); } }, Some(AssistantSettingsContentInner::Legacy(settings)) => { settings.dock = Some(dock); } None => { self.inner = Some(AssistantSettingsContentInner::for_v2( AssistantSettingsContentV2 { dock: Some(dock), ..Default::default() }, )) } } } pub fn set_model(&mut self, language_model: Arc) { let model = language_model.id().0.to_string(); let provider = language_model.provider_id().0.to_string(); match &mut self.inner { Some(AssistantSettingsContentInner::Versioned(settings)) => match **settings { VersionedAssistantSettingsContent::V1(ref mut settings) => { match provider.as_ref() { "zed.dev" => { log::warn!("attempted to set zed.dev model on outdated settings"); } "anthropic" => { let api_url = match &settings.provider { Some(AssistantProviderContentV1::Anthropic { api_url, .. }) => { api_url.clone() } _ => None, }; settings.provider = Some(AssistantProviderContentV1::Anthropic { default_model: AnthropicModel::from_id(&model).ok(), api_url, }); } "ollama" => { let api_url = match &settings.provider { Some(AssistantProviderContentV1::Ollama { api_url, .. }) => { api_url.clone() } _ => None, }; settings.provider = Some(AssistantProviderContentV1::Ollama { default_model: Some(ollama::Model::new(&model, None, None)), api_url, }); } "lmstudio" => { let api_url = match &settings.provider { Some(AssistantProviderContentV1::LmStudio { api_url, .. }) => { api_url.clone() } _ => None, }; settings.provider = Some(AssistantProviderContentV1::LmStudio { default_model: Some(lmstudio::Model::new(&model, None, None)), api_url, }); } "openai" => { let (api_url, available_models) = match &settings.provider { Some(AssistantProviderContentV1::OpenAi { api_url, available_models, .. }) => (api_url.clone(), available_models.clone()), _ => (None, None), }; settings.provider = Some(AssistantProviderContentV1::OpenAi { default_model: OpenAiModel::from_id(&model).ok(), api_url, available_models, }); } "deepseek" => { let api_url = match &settings.provider { Some(AssistantProviderContentV1::DeepSeek { api_url, .. }) => { api_url.clone() } _ => None, }; settings.provider = Some(AssistantProviderContentV1::DeepSeek { default_model: DeepseekModel::from_id(&model).ok(), api_url, }); } _ => {} } } VersionedAssistantSettingsContent::V2(ref mut settings) => { settings.default_model = Some(LanguageModelSelection { provider, model }); } }, Some(AssistantSettingsContentInner::Legacy(settings)) => { if let Ok(model) = OpenAiModel::from_id(&language_model.id().0) { settings.default_open_ai_model = Some(model); } } None => { self.inner = Some(AssistantSettingsContentInner::for_v2( AssistantSettingsContentV2 { default_model: Some(LanguageModelSelection { provider, model }), ..Default::default() }, )); } } } pub fn set_inline_assistant_model(&mut self, provider: String, model: String) { self.v2_setting(|setting| { setting.inline_assistant_model = Some(LanguageModelSelection { provider, model }); Ok(()) }) .ok(); } pub fn set_commit_message_model(&mut self, provider: String, model: String) { self.v2_setting(|setting| { setting.commit_message_model = Some(LanguageModelSelection { provider, model }); Ok(()) }) .ok(); } pub fn v2_setting( &mut self, f: impl FnOnce(&mut AssistantSettingsContentV2) -> anyhow::Result<()>, ) -> anyhow::Result<()> { match self.inner.get_or_insert_with(|| { AssistantSettingsContentInner::for_v2(AssistantSettingsContentV2 { ..Default::default() }) }) { AssistantSettingsContentInner::Versioned(boxed) => { if let VersionedAssistantSettingsContent::V2(ref mut settings) = **boxed { f(settings) } else { Ok(()) } } _ => Ok(()), } } pub fn set_thread_summary_model(&mut self, provider: String, model: String) { self.v2_setting(|setting| { setting.thread_summary_model = Some(LanguageModelSelection { provider, model }); Ok(()) }) .ok(); } pub fn set_always_allow_tool_actions(&mut self, allow: bool) { self.v2_setting(|setting| { setting.always_allow_tool_actions = Some(allow); Ok(()) }) .ok(); } pub fn set_profile(&mut self, profile_id: AgentProfileId) { self.v2_setting(|setting| { setting.default_profile = Some(profile_id); Ok(()) }) .ok(); } pub fn create_profile( &mut self, profile_id: AgentProfileId, profile: AgentProfile, ) -> Result<()> { self.v2_setting(|settings| { let profiles = settings.profiles.get_or_insert_default(); if profiles.contains_key(&profile_id) { bail!("profile with ID '{profile_id}' already exists"); } profiles.insert( profile_id, AgentProfileContent { name: profile.name.into(), tools: profile.tools, enable_all_context_servers: Some(profile.enable_all_context_servers), context_servers: profile .context_servers .into_iter() .map(|(server_id, preset)| { ( server_id, ContextServerPresetContent { tools: preset.tools, }, ) }) .collect(), }, ); Ok(()) }) } } #[derive(Clone, Serialize, Deserialize, JsonSchema, Debug)] #[serde(tag = "version")] pub enum VersionedAssistantSettingsContent { #[serde(rename = "1")] V1(AssistantSettingsContentV1), #[serde(rename = "2")] V2(AssistantSettingsContentV2), } impl Default for VersionedAssistantSettingsContent { fn default() -> Self { Self::V2(AssistantSettingsContentV2 { enabled: None, button: None, dock: None, default_width: None, default_height: None, default_model: None, inline_assistant_model: None, commit_message_model: None, thread_summary_model: None, inline_alternatives: None, enable_experimental_live_diffs: None, default_profile: None, profiles: None, always_allow_tool_actions: None, notify_when_agent_waiting: None, stream_edits: None, single_file_review: None, }) } } #[derive(Clone, Serialize, Deserialize, JsonSchema, Debug, Default)] pub struct AssistantSettingsContentV2 { /// Whether the Assistant is enabled. /// /// Default: true enabled: Option, /// Whether to show the assistant panel button in the status bar. /// /// Default: true button: Option, /// Where to dock the assistant. /// /// Default: right dock: Option, /// Default width in pixels when the assistant is docked to the left or right. /// /// Default: 640 default_width: Option, /// Default height in pixels when the assistant is docked to the bottom. /// /// Default: 320 default_height: Option, /// The default model to use when creating new chats and for other features when a specific model is not specified. default_model: Option, /// Model to use for the inline assistant. Defaults to default_model when not specified. inline_assistant_model: Option, /// Model to use for generating git commit messages. Defaults to default_model when not specified. commit_message_model: Option, /// Model to use for generating thread summaries. Defaults to default_model when not specified. thread_summary_model: Option, /// Additional models with which to generate alternatives when performing inline assists. inline_alternatives: Option>, /// Enable experimental live diffs in the assistant panel. /// /// Default: false enable_experimental_live_diffs: Option, /// The default profile to use in the Agent. /// /// Default: write default_profile: Option, /// The available agent profiles. pub profiles: Option>, /// Whenever a tool action would normally wait for your confirmation /// that you allow it, always choose to allow it. /// /// Default: false always_allow_tool_actions: Option, /// Where to show a popup notification when the agent is waiting for user input. /// /// Default: "primary_screen" notify_when_agent_waiting: Option, /// Whether to stream edits from the agent as they are received. /// /// Default: false stream_edits: Option, /// Whether to display agent edits in single-file editors in addition to the review multibuffer pane. /// /// Default: true single_file_review: Option, } #[derive(Clone, Debug, Serialize, Deserialize, JsonSchema, PartialEq)] pub struct LanguageModelSelection { #[schemars(schema_with = "providers_schema")] pub provider: String, pub model: String, } fn providers_schema(_: &mut schemars::r#gen::SchemaGenerator) -> schemars::schema::Schema { schemars::schema::SchemaObject { enum_values: Some(vec![ "anthropic".into(), "bedrock".into(), "google".into(), "lmstudio".into(), "ollama".into(), "openai".into(), "zed.dev".into(), "copilot_chat".into(), "deepseek".into(), ]), ..Default::default() } .into() } impl Default for LanguageModelSelection { fn default() -> Self { Self { provider: "openai".to_string(), model: "gpt-4".to_string(), } } } #[derive(Debug, PartialEq, Clone, Serialize, Deserialize, JsonSchema)] pub struct AgentProfileContent { pub name: Arc, #[serde(default)] pub tools: IndexMap, bool>, /// Whether all context servers are enabled by default. pub enable_all_context_servers: Option, #[serde(default)] pub context_servers: IndexMap, ContextServerPresetContent>, } #[derive(Debug, PartialEq, Clone, Default, Serialize, Deserialize, JsonSchema)] pub struct ContextServerPresetContent { pub tools: IndexMap, bool>, } #[derive(Clone, Serialize, Deserialize, JsonSchema, Debug)] pub struct AssistantSettingsContentV1 { /// Whether the Assistant is enabled. /// /// Default: true enabled: Option, /// Whether to show the assistant panel button in the status bar. /// /// Default: true button: Option, /// Where to dock the assistant. /// /// Default: right dock: Option, /// Default width in pixels when the assistant is docked to the left or right. /// /// Default: 640 default_width: Option, /// Default height in pixels when the assistant is docked to the bottom. /// /// Default: 320 default_height: Option, /// The provider of the assistant service. /// /// This can be "openai", "anthropic", "ollama", "lmstudio", "deepseek", "zed.dev" /// each with their respective default models and configurations. provider: Option, } #[derive(Clone, Serialize, Deserialize, JsonSchema, Debug)] pub struct LegacyAssistantSettingsContent { /// Whether to show the assistant panel button in the status bar. /// /// Default: true pub button: Option, /// Where to dock the assistant. /// /// Default: right pub dock: Option, /// Default width in pixels when the assistant is docked to the left or right. /// /// Default: 640 pub default_width: Option, /// Default height in pixels when the assistant is docked to the bottom. /// /// Default: 320 pub default_height: Option, /// The default OpenAI model to use when creating new chats. /// /// Default: gpt-4-1106-preview pub default_open_ai_model: Option, /// OpenAI API base URL to use when creating new chats. /// /// Default: pub openai_api_url: Option, } impl Settings for AssistantSettings { const KEY: Option<&'static str> = Some("assistant"); const PRESERVED_KEYS: Option<&'static [&'static str]> = Some(&["version"]); type FileContent = AssistantSettingsContent; fn load( sources: SettingsSources, _: &mut gpui::App, ) -> anyhow::Result { let mut settings = AssistantSettings::default(); for value in sources.defaults_and_customizations() { if value.is_version_outdated() { settings.using_outdated_settings_version = true; } let value = value.upgrade(); merge(&mut settings.enabled, value.enabled); merge(&mut settings.button, value.button); merge(&mut settings.dock, value.dock); merge( &mut settings.default_width, value.default_width.map(Into::into), ); merge( &mut settings.default_height, value.default_height.map(Into::into), ); merge(&mut settings.default_model, value.default_model); settings.inline_assistant_model = value .inline_assistant_model .or(settings.inline_assistant_model.take()); settings.commit_message_model = value .commit_message_model .or(settings.commit_message_model.take()); settings.thread_summary_model = value .thread_summary_model .or(settings.thread_summary_model.take()); merge(&mut settings.inline_alternatives, value.inline_alternatives); merge( &mut settings.enable_experimental_live_diffs, value.enable_experimental_live_diffs, ); merge( &mut settings.always_allow_tool_actions, value.always_allow_tool_actions, ); merge( &mut settings.notify_when_agent_waiting, value.notify_when_agent_waiting, ); merge(&mut settings.stream_edits, value.stream_edits); merge(&mut settings.single_file_review, value.single_file_review); merge(&mut settings.default_profile, value.default_profile); if let Some(profiles) = value.profiles { settings .profiles .extend(profiles.into_iter().map(|(id, profile)| { ( id, AgentProfile { name: profile.name.into(), tools: profile.tools, enable_all_context_servers: profile .enable_all_context_servers .unwrap_or_default(), context_servers: profile .context_servers .into_iter() .map(|(context_server_id, preset)| { ( context_server_id, ContextServerPreset { tools: preset.tools.clone(), }, ) }) .collect(), }, ) })); } } Ok(settings) } fn import_from_vscode(vscode: &settings::VsCodeSettings, current: &mut Self::FileContent) { if let Some(b) = vscode .read_value("chat.agent.enabled") .and_then(|b| b.as_bool()) { match &mut current.inner { Some(AssistantSettingsContentInner::Versioned(versioned)) => { match versioned.as_mut() { VersionedAssistantSettingsContent::V1(setting) => { setting.enabled = Some(b); setting.button = Some(b); } VersionedAssistantSettingsContent::V2(setting) => { setting.enabled = Some(b); setting.button = Some(b); } } } Some(AssistantSettingsContentInner::Legacy(setting)) => setting.button = Some(b), None => { current.inner = Some(AssistantSettingsContentInner::for_v2( AssistantSettingsContentV2 { enabled: Some(b), button: Some(b), ..Default::default() }, )); } } } } } fn merge(target: &mut T, value: Option) { if let Some(value) = value { *target = value; } } #[cfg(test)] mod tests { use fs::Fs; use gpui::{ReadGlobal, TestAppContext}; use super::*; #[gpui::test] async fn test_deserialize_assistant_settings_with_version(cx: &mut TestAppContext) { let fs = fs::FakeFs::new(cx.executor().clone()); fs.create_dir(paths::settings_file().parent().unwrap()) .await .unwrap(); cx.update(|cx| { let test_settings = settings::SettingsStore::test(cx); cx.set_global(test_settings); AssistantSettings::register(cx); }); cx.update(|cx| { assert!(!AssistantSettings::get_global(cx).using_outdated_settings_version); assert_eq!( AssistantSettings::get_global(cx).default_model, LanguageModelSelection { provider: "zed.dev".into(), model: "claude-3-7-sonnet-latest".into(), } ); }); cx.update(|cx| { settings::SettingsStore::global(cx).update_settings_file::( fs.clone(), |settings, _| { *settings = AssistantSettingsContent { inner: Some(AssistantSettingsContentInner::for_v2( AssistantSettingsContentV2 { default_model: Some(LanguageModelSelection { provider: "test-provider".into(), model: "gpt-99".into(), }), inline_assistant_model: None, commit_message_model: None, thread_summary_model: None, inline_alternatives: None, enabled: None, button: None, dock: None, default_width: None, default_height: None, enable_experimental_live_diffs: None, default_profile: None, profiles: None, always_allow_tool_actions: None, notify_when_agent_waiting: None, stream_edits: None, single_file_review: None, }, )), } }, ); }); cx.run_until_parked(); let raw_settings_value = fs.load(paths::settings_file()).await.unwrap(); assert!(raw_settings_value.contains(r#""version": "2""#)); #[derive(Debug, Deserialize)] struct AssistantSettingsTest { assistant: AssistantSettingsContent, } let assistant_settings: AssistantSettingsTest = serde_json_lenient::from_str(&raw_settings_value).unwrap(); assert!(!assistant_settings.assistant.is_version_outdated()); } }