mod agent_profile; use std::sync::Arc; use ::open_ai::Model as OpenAiModel; use anthropic::Model as AnthropicModel; use anyhow::{Result, bail}; use deepseek::Model as DeepseekModel; use feature_flags::{Assistant2FeatureFlag, FeatureFlagAppExt}; use gpui::{App, Pixels}; use indexmap::IndexMap; use language_model::{CloudModel, LanguageModel}; use lmstudio::Model as LmStudioModel; use ollama::Model as OllamaModel; use schemars::{JsonSchema, schema::Schema}; use serde::{Deserialize, Serialize}; use settings::{Settings, SettingsSources}; pub use crate::agent_profile::*; #[derive(Copy, Clone, Default, Debug, Serialize, Deserialize, JsonSchema)] #[serde(rename_all = "snake_case")] pub enum AssistantDockPosition { Left, #[default] Right, Bottom, } #[derive(Copy, Clone, Default, Debug, Serialize, Deserialize, JsonSchema, PartialEq, Eq)] #[serde(rename_all = "snake_case")] pub enum NotifyWhenAgentWaiting { #[default] PrimaryScreen, AllScreens, Never, } #[derive(Clone, Debug, Serialize, Deserialize, JsonSchema, PartialEq)] #[serde(tag = "name", rename_all = "snake_case")] pub enum AssistantProviderContentV1 { #[serde(rename = "zed.dev")] ZedDotDev { default_model: Option }, #[serde(rename = "openai")] OpenAi { default_model: Option, api_url: Option, available_models: Option>, }, #[serde(rename = "anthropic")] Anthropic { default_model: Option, api_url: Option, }, #[serde(rename = "ollama")] Ollama { default_model: Option, api_url: Option, }, #[serde(rename = "lmstudio")] LmStudio { default_model: Option, api_url: Option, }, #[serde(rename = "deepseek")] DeepSeek { default_model: Option, api_url: Option, }, } #[derive(Debug, Default)] pub struct AssistantSettings { pub enabled: bool, pub button: bool, pub dock: AssistantDockPosition, pub default_width: Pixels, pub default_height: Pixels, pub default_model: LanguageModelSelection, pub editor_model: LanguageModelSelection, pub inline_alternatives: Vec, pub using_outdated_settings_version: bool, pub enable_experimental_live_diffs: bool, pub default_profile: AgentProfileId, pub profiles: IndexMap, pub always_allow_tool_actions: bool, pub notify_when_agent_waiting: NotifyWhenAgentWaiting, } impl AssistantSettings { pub fn are_live_diffs_enabled(&self, cx: &App) -> bool { if cx.has_flag::() { return false; } cx.is_staff() || self.enable_experimental_live_diffs } } /// Assistant panel settings #[derive(Clone, Serialize, Deserialize, Debug)] #[serde(untagged)] pub enum AssistantSettingsContent { Versioned(VersionedAssistantSettingsContent), Legacy(LegacyAssistantSettingsContent), } impl JsonSchema for AssistantSettingsContent { fn schema_name() -> String { VersionedAssistantSettingsContent::schema_name() } fn json_schema(r#gen: &mut schemars::r#gen::SchemaGenerator) -> Schema { VersionedAssistantSettingsContent::json_schema(r#gen) } fn is_referenceable() -> bool { VersionedAssistantSettingsContent::is_referenceable() } } impl Default for AssistantSettingsContent { fn default() -> Self { Self::Versioned(VersionedAssistantSettingsContent::default()) } } impl AssistantSettingsContent { pub fn is_version_outdated(&self) -> bool { match self { AssistantSettingsContent::Versioned(settings) => match settings { VersionedAssistantSettingsContent::V1(_) => true, VersionedAssistantSettingsContent::V2(_) => false, }, AssistantSettingsContent::Legacy(_) => true, } } fn upgrade(&self) -> AssistantSettingsContentV2 { match self { AssistantSettingsContent::Versioned(settings) => match settings { VersionedAssistantSettingsContent::V1(settings) => AssistantSettingsContentV2 { enabled: settings.enabled, button: settings.button, dock: settings.dock, default_width: settings.default_width, default_height: settings.default_width, default_model: settings .provider .clone() .and_then(|provider| match provider { AssistantProviderContentV1::ZedDotDev { default_model } => { default_model.map(|model| LanguageModelSelection { provider: "zed.dev".to_string(), model: model.id().to_string(), }) } AssistantProviderContentV1::OpenAi { default_model, .. } => { default_model.map(|model| LanguageModelSelection { provider: "openai".to_string(), model: model.id().to_string(), }) } AssistantProviderContentV1::Anthropic { default_model, .. } => { default_model.map(|model| LanguageModelSelection { provider: "anthropic".to_string(), model: model.id().to_string(), }) } AssistantProviderContentV1::Ollama { default_model, .. } => { default_model.map(|model| LanguageModelSelection { provider: "ollama".to_string(), model: model.id().to_string(), }) } AssistantProviderContentV1::LmStudio { default_model, .. } => { default_model.map(|model| LanguageModelSelection { provider: "lmstudio".to_string(), model: model.id().to_string(), }) } AssistantProviderContentV1::DeepSeek { default_model, .. } => { default_model.map(|model| LanguageModelSelection { provider: "deepseek".to_string(), model: model.id().to_string(), }) } }), editor_model: None, inline_alternatives: None, enable_experimental_live_diffs: None, default_profile: None, profiles: None, always_allow_tool_actions: None, notify_when_agent_waiting: None, }, VersionedAssistantSettingsContent::V2(settings) => settings.clone(), }, AssistantSettingsContent::Legacy(settings) => AssistantSettingsContentV2 { enabled: None, button: settings.button, dock: settings.dock, default_width: settings.default_width, default_height: settings.default_height, default_model: Some(LanguageModelSelection { provider: "openai".to_string(), model: settings .default_open_ai_model .clone() .unwrap_or_default() .id() .to_string(), }), editor_model: None, inline_alternatives: None, enable_experimental_live_diffs: None, default_profile: None, profiles: None, always_allow_tool_actions: None, notify_when_agent_waiting: None, }, } } pub fn set_dock(&mut self, dock: AssistantDockPosition) { match self { AssistantSettingsContent::Versioned(settings) => match settings { VersionedAssistantSettingsContent::V1(settings) => { settings.dock = Some(dock); } VersionedAssistantSettingsContent::V2(settings) => { settings.dock = Some(dock); } }, AssistantSettingsContent::Legacy(settings) => { settings.dock = Some(dock); } } } pub fn set_model(&mut self, language_model: Arc) { let model = language_model.id().0.to_string(); let provider = language_model.provider_id().0.to_string(); match self { AssistantSettingsContent::Versioned(settings) => match settings { VersionedAssistantSettingsContent::V1(settings) => match provider.as_ref() { "zed.dev" => { log::warn!("attempted to set zed.dev model on outdated settings"); } "anthropic" => { let api_url = match &settings.provider { Some(AssistantProviderContentV1::Anthropic { api_url, .. }) => { api_url.clone() } _ => None, }; settings.provider = Some(AssistantProviderContentV1::Anthropic { default_model: AnthropicModel::from_id(&model).ok(), api_url, }); } "ollama" => { let api_url = match &settings.provider { Some(AssistantProviderContentV1::Ollama { api_url, .. }) => { api_url.clone() } _ => None, }; settings.provider = Some(AssistantProviderContentV1::Ollama { default_model: Some(ollama::Model::new(&model, None, None)), api_url, }); } "lmstudio" => { let api_url = match &settings.provider { Some(AssistantProviderContentV1::LmStudio { api_url, .. }) => { api_url.clone() } _ => None, }; settings.provider = Some(AssistantProviderContentV1::LmStudio { default_model: Some(lmstudio::Model::new(&model, None, None)), api_url, }); } "openai" => { let (api_url, available_models) = match &settings.provider { Some(AssistantProviderContentV1::OpenAi { api_url, available_models, .. }) => (api_url.clone(), available_models.clone()), _ => (None, None), }; settings.provider = Some(AssistantProviderContentV1::OpenAi { default_model: OpenAiModel::from_id(&model).ok(), api_url, available_models, }); } "deepseek" => { let api_url = match &settings.provider { Some(AssistantProviderContentV1::DeepSeek { api_url, .. }) => { api_url.clone() } _ => None, }; settings.provider = Some(AssistantProviderContentV1::DeepSeek { default_model: DeepseekModel::from_id(&model).ok(), api_url, }); } _ => {} }, VersionedAssistantSettingsContent::V2(settings) => { settings.default_model = Some(LanguageModelSelection { provider, model }); } }, AssistantSettingsContent::Legacy(settings) => { if let Ok(model) = OpenAiModel::from_id(&language_model.id().0) { settings.default_open_ai_model = Some(model); } } } } pub fn set_always_allow_tool_actions(&mut self, allow: bool) { let AssistantSettingsContent::Versioned(VersionedAssistantSettingsContent::V2(settings)) = self else { return; }; settings.always_allow_tool_actions = Some(allow); } pub fn set_profile(&mut self, profile_id: AgentProfileId) { let AssistantSettingsContent::Versioned(VersionedAssistantSettingsContent::V2(settings)) = self else { return; }; settings.default_profile = Some(profile_id); } pub fn create_profile( &mut self, profile_id: AgentProfileId, profile: AgentProfile, ) -> Result<()> { let AssistantSettingsContent::Versioned(VersionedAssistantSettingsContent::V2(settings)) = self else { return Ok(()); }; let profiles = settings.profiles.get_or_insert_default(); if profiles.contains_key(&profile_id) { bail!("profile with ID '{profile_id}' already exists"); } profiles.insert( profile_id, AgentProfileContent { name: profile.name.into(), tools: profile.tools, enable_all_context_servers: Some(profile.enable_all_context_servers), context_servers: profile .context_servers .into_iter() .map(|(server_id, preset)| { ( server_id, ContextServerPresetContent { tools: preset.tools, }, ) }) .collect(), }, ); Ok(()) } } #[derive(Clone, Serialize, Deserialize, JsonSchema, Debug)] #[serde(tag = "version")] pub enum VersionedAssistantSettingsContent { #[serde(rename = "1")] V1(AssistantSettingsContentV1), #[serde(rename = "2")] V2(AssistantSettingsContentV2), } impl Default for VersionedAssistantSettingsContent { fn default() -> Self { Self::V2(AssistantSettingsContentV2 { enabled: None, button: None, dock: None, default_width: None, default_height: None, default_model: None, editor_model: None, inline_alternatives: None, enable_experimental_live_diffs: None, default_profile: None, profiles: None, always_allow_tool_actions: None, notify_when_agent_waiting: None, }) } } #[derive(Clone, Serialize, Deserialize, JsonSchema, Debug)] pub struct AssistantSettingsContentV2 { /// Whether the Assistant is enabled. /// /// Default: true enabled: Option, /// Whether to show the assistant panel button in the status bar. /// /// Default: true button: Option, /// Where to dock the assistant. /// /// Default: right dock: Option, /// Default width in pixels when the assistant is docked to the left or right. /// /// Default: 640 default_width: Option, /// Default height in pixels when the assistant is docked to the bottom. /// /// Default: 320 default_height: Option, /// The default model to use when creating new chats. default_model: Option, /// The model to use when applying edits from the assistant. editor_model: Option, /// Additional models with which to generate alternatives when performing inline assists. inline_alternatives: Option>, /// Enable experimental live diffs in the assistant panel. /// /// Default: false enable_experimental_live_diffs: Option, /// The default profile to use in the Agent. /// /// Default: write default_profile: Option, /// The available agent profiles. pub profiles: Option>, /// Whenever a tool action would normally wait for your confirmation /// that you allow it, always choose to allow it. /// /// Default: false always_allow_tool_actions: Option, /// Where to show a popup notification when the agent is waiting for user input. /// /// Default: "primary_screen" notify_when_agent_waiting: Option, } #[derive(Clone, Debug, Serialize, Deserialize, JsonSchema, PartialEq)] pub struct LanguageModelSelection { #[schemars(schema_with = "providers_schema")] pub provider: String, pub model: String, } fn providers_schema(_: &mut schemars::r#gen::SchemaGenerator) -> schemars::schema::Schema { schemars::schema::SchemaObject { enum_values: Some(vec![ "anthropic".into(), "bedrock".into(), "google".into(), "lmstudio".into(), "ollama".into(), "openai".into(), "zed.dev".into(), "copilot_chat".into(), "deepseek".into(), ]), ..Default::default() } .into() } impl Default for LanguageModelSelection { fn default() -> Self { Self { provider: "openai".to_string(), model: "gpt-4".to_string(), } } } #[derive(Debug, PartialEq, Clone, Serialize, Deserialize, JsonSchema)] pub struct AgentProfileContent { pub name: Arc, #[serde(default)] pub tools: IndexMap, bool>, /// Whether all context servers are enabled by default. pub enable_all_context_servers: Option, #[serde(default)] pub context_servers: IndexMap, ContextServerPresetContent>, } #[derive(Debug, PartialEq, Clone, Default, Serialize, Deserialize, JsonSchema)] pub struct ContextServerPresetContent { pub tools: IndexMap, bool>, } #[derive(Clone, Serialize, Deserialize, JsonSchema, Debug)] pub struct AssistantSettingsContentV1 { /// Whether the Assistant is enabled. /// /// Default: true enabled: Option, /// Whether to show the assistant panel button in the status bar. /// /// Default: true button: Option, /// Where to dock the assistant. /// /// Default: right dock: Option, /// Default width in pixels when the assistant is docked to the left or right. /// /// Default: 640 default_width: Option, /// Default height in pixels when the assistant is docked to the bottom. /// /// Default: 320 default_height: Option, /// The provider of the assistant service. /// /// This can be "openai", "anthropic", "ollama", "lmstudio", "deepseek", "zed.dev" /// each with their respective default models and configurations. provider: Option, } #[derive(Clone, Serialize, Deserialize, JsonSchema, Debug)] pub struct LegacyAssistantSettingsContent { /// Whether to show the assistant panel button in the status bar. /// /// Default: true pub button: Option, /// Where to dock the assistant. /// /// Default: right pub dock: Option, /// Default width in pixels when the assistant is docked to the left or right. /// /// Default: 640 pub default_width: Option, /// Default height in pixels when the assistant is docked to the bottom. /// /// Default: 320 pub default_height: Option, /// The default OpenAI model to use when creating new chats. /// /// Default: gpt-4-1106-preview pub default_open_ai_model: Option, /// OpenAI API base URL to use when creating new chats. /// /// Default: pub openai_api_url: Option, } impl Settings for AssistantSettings { const KEY: Option<&'static str> = Some("assistant"); const PRESERVED_KEYS: Option<&'static [&'static str]> = Some(&["version"]); type FileContent = AssistantSettingsContent; fn load( sources: SettingsSources, _: &mut gpui::App, ) -> anyhow::Result { let mut settings = AssistantSettings::default(); for value in sources.defaults_and_customizations() { if value.is_version_outdated() { settings.using_outdated_settings_version = true; } let value = value.upgrade(); merge(&mut settings.enabled, value.enabled); merge(&mut settings.button, value.button); merge(&mut settings.dock, value.dock); merge( &mut settings.default_width, value.default_width.map(Into::into), ); merge( &mut settings.default_height, value.default_height.map(Into::into), ); merge(&mut settings.default_model, value.default_model); merge(&mut settings.editor_model, value.editor_model); merge(&mut settings.inline_alternatives, value.inline_alternatives); merge( &mut settings.enable_experimental_live_diffs, value.enable_experimental_live_diffs, ); merge( &mut settings.always_allow_tool_actions, value.always_allow_tool_actions, ); merge( &mut settings.notify_when_agent_waiting, value.notify_when_agent_waiting, ); merge(&mut settings.default_profile, value.default_profile); if let Some(profiles) = value.profiles { settings .profiles .extend(profiles.into_iter().map(|(id, profile)| { ( id, AgentProfile { name: profile.name.into(), tools: profile.tools, enable_all_context_servers: profile .enable_all_context_servers .unwrap_or_default(), context_servers: profile .context_servers .into_iter() .map(|(context_server_id, preset)| { ( context_server_id, ContextServerPreset { tools: preset.tools.clone(), }, ) }) .collect(), }, ) })); } } Ok(settings) } } fn merge(target: &mut T, value: Option) { if let Some(value) = value { *target = value; } } #[cfg(test)] mod tests { use fs::Fs; use gpui::{ReadGlobal, TestAppContext}; use super::*; #[gpui::test] async fn test_deserialize_assistant_settings_with_version(cx: &mut TestAppContext) { let fs = fs::FakeFs::new(cx.executor().clone()); fs.create_dir(paths::settings_file().parent().unwrap()) .await .unwrap(); cx.update(|cx| { let test_settings = settings::SettingsStore::test(cx); cx.set_global(test_settings); AssistantSettings::register(cx); }); cx.update(|cx| { assert!(!AssistantSettings::get_global(cx).using_outdated_settings_version); assert_eq!( AssistantSettings::get_global(cx).default_model, LanguageModelSelection { provider: "zed.dev".into(), model: "claude-3-5-sonnet-latest".into(), } ); }); cx.update(|cx| { settings::SettingsStore::global(cx).update_settings_file::( fs.clone(), |settings, _| { *settings = AssistantSettingsContent::Versioned( VersionedAssistantSettingsContent::V2(AssistantSettingsContentV2 { default_model: Some(LanguageModelSelection { provider: "test-provider".into(), model: "gpt-99".into(), }), editor_model: Some(LanguageModelSelection { provider: "test-provider".into(), model: "gpt-99".into(), }), inline_alternatives: None, enabled: None, button: None, dock: None, default_width: None, default_height: None, enable_experimental_live_diffs: None, default_profile: None, profiles: None, always_allow_tool_actions: None, notify_when_agent_waiting: None, }), ) }, ); }); cx.run_until_parked(); let raw_settings_value = fs.load(paths::settings_file()).await.unwrap(); assert!(raw_settings_value.contains(r#""version": "2""#)); #[derive(Debug, Deserialize)] struct AssistantSettingsTest { assistant: AssistantSettingsContent, } let assistant_settings: AssistantSettingsTest = serde_json_lenient::from_str(&raw_settings_value).unwrap(); assert!(!assistant_settings.assistant.is_version_outdated()); } }