pub mod extension; pub mod registry; use std::{path::Path, sync::Arc}; use anyhow::{Context as _, Result}; use collections::{HashMap, HashSet}; use context_server::{ContextServer, ContextServerCommand, ContextServerId}; use futures::{FutureExt as _, future::join_all}; use gpui::{App, AsyncApp, Context, Entity, EventEmitter, Subscription, Task, WeakEntity, actions}; use registry::ContextServerDescriptorRegistry; use settings::{Settings as _, SettingsStore}; use util::ResultExt as _; use crate::{ Project, project_settings::{ContextServerSettings, ProjectSettings}, worktree_store::WorktreeStore, }; pub fn init(cx: &mut App) { extension::init(cx); } actions!( context_server, [ /// Restarts the context server. Restart ] ); #[derive(Debug, Clone, PartialEq, Eq, Hash)] pub enum ContextServerStatus { Starting, Running, Stopped, Error(Arc), } impl ContextServerStatus { fn from_state(state: &ContextServerState) -> Self { match state { ContextServerState::Starting { .. } => ContextServerStatus::Starting, ContextServerState::Running { .. } => ContextServerStatus::Running, ContextServerState::Stopped { .. } => ContextServerStatus::Stopped, ContextServerState::Error { error, .. } => ContextServerStatus::Error(error.clone()), } } } enum ContextServerState { Starting { server: Arc, configuration: Arc, _task: Task<()>, }, Running { server: Arc, configuration: Arc, }, Stopped { server: Arc, configuration: Arc, }, Error { server: Arc, configuration: Arc, error: Arc, }, } impl ContextServerState { pub fn server(&self) -> Arc { match self { ContextServerState::Starting { server, .. } => server.clone(), ContextServerState::Running { server, .. } => server.clone(), ContextServerState::Stopped { server, .. } => server.clone(), ContextServerState::Error { server, .. } => server.clone(), } } pub fn configuration(&self) -> Arc { match self { ContextServerState::Starting { configuration, .. } => configuration.clone(), ContextServerState::Running { configuration, .. } => configuration.clone(), ContextServerState::Stopped { configuration, .. } => configuration.clone(), ContextServerState::Error { configuration, .. } => configuration.clone(), } } } #[derive(Debug, PartialEq, Eq)] pub enum ContextServerConfiguration { Custom { command: ContextServerCommand, }, Extension { command: ContextServerCommand, settings: serde_json::Value, }, } impl ContextServerConfiguration { pub fn command(&self) -> &ContextServerCommand { match self { ContextServerConfiguration::Custom { command } => command, ContextServerConfiguration::Extension { command, .. } => command, } } pub async fn from_settings( settings: ContextServerSettings, id: ContextServerId, registry: Entity, worktree_store: Entity, cx: &AsyncApp, ) -> Option { match settings { ContextServerSettings::Custom { enabled: _, command, } => Some(ContextServerConfiguration::Custom { command }), ContextServerSettings::Extension { enabled: _, settings, } => { let descriptor = cx .update(|cx| registry.read(cx).context_server_descriptor(&id.0)) .ok() .flatten()?; let command = descriptor.command(worktree_store, cx).await.log_err()?; Some(ContextServerConfiguration::Extension { command, settings }) } } } } pub type ContextServerFactory = Box) -> Arc>; pub struct ContextServerStore { context_server_settings: HashMap, ContextServerSettings>, servers: HashMap, worktree_store: Entity, project: WeakEntity, registry: Entity, update_servers_task: Option>>, context_server_factory: Option, needs_server_update: bool, _subscriptions: Vec, } pub enum Event { ServerStatusChanged { server_id: ContextServerId, status: ContextServerStatus, }, } impl EventEmitter for ContextServerStore {} impl ContextServerStore { pub fn new( worktree_store: Entity, weak_project: WeakEntity, cx: &mut Context, ) -> Self { Self::new_internal( true, None, ContextServerDescriptorRegistry::default_global(cx), worktree_store, weak_project, cx, ) } /// Returns all configured context server ids, regardless of enabled state. pub fn configured_server_ids(&self) -> Vec { self.context_server_settings .keys() .cloned() .map(ContextServerId) .collect() } #[cfg(any(test, feature = "test-support"))] pub fn test( registry: Entity, worktree_store: Entity, weak_project: WeakEntity, cx: &mut Context, ) -> Self { Self::new_internal(false, None, registry, worktree_store, weak_project, cx) } #[cfg(any(test, feature = "test-support"))] pub fn test_maintain_server_loop( context_server_factory: ContextServerFactory, registry: Entity, worktree_store: Entity, weak_project: WeakEntity, cx: &mut Context, ) -> Self { Self::new_internal( true, Some(context_server_factory), registry, worktree_store, weak_project, cx, ) } fn new_internal( maintain_server_loop: bool, context_server_factory: Option, registry: Entity, worktree_store: Entity, weak_project: WeakEntity, cx: &mut Context, ) -> Self { let subscriptions = if maintain_server_loop { vec![ cx.observe(®istry, |this, _registry, cx| { this.available_context_servers_changed(cx); }), cx.observe_global::(|this, cx| { let settings = Self::resolve_context_server_settings(&this.worktree_store, cx); if &this.context_server_settings == settings { return; } this.context_server_settings = settings.clone(); this.available_context_servers_changed(cx); }), ] } else { Vec::new() }; let mut this = Self { _subscriptions: subscriptions, context_server_settings: Self::resolve_context_server_settings(&worktree_store, cx) .clone(), worktree_store, project: weak_project, registry, needs_server_update: false, servers: HashMap::default(), update_servers_task: None, context_server_factory, }; if maintain_server_loop { this.available_context_servers_changed(cx); } this } pub fn get_server(&self, id: &ContextServerId) -> Option> { self.servers.get(id).map(|state| state.server()) } pub fn get_running_server(&self, id: &ContextServerId) -> Option> { if let Some(ContextServerState::Running { server, .. }) = self.servers.get(id) { Some(server.clone()) } else { None } } pub fn status_for_server(&self, id: &ContextServerId) -> Option { self.servers.get(id).map(ContextServerStatus::from_state) } pub fn configuration_for_server( &self, id: &ContextServerId, ) -> Option> { self.servers.get(id).map(|state| state.configuration()) } pub fn all_server_ids(&self) -> Vec { self.servers.keys().cloned().collect() } pub fn running_servers(&self) -> Vec> { self.servers .values() .filter_map(|state| { if let ContextServerState::Running { server, .. } = state { Some(server.clone()) } else { None } }) .collect() } pub fn start_server(&mut self, server: Arc, cx: &mut Context) { cx.spawn(async move |this, cx| { let this = this.upgrade().context("Context server store dropped")?; let settings = this .update(cx, |this, _| { this.context_server_settings.get(&server.id().0).cloned() }) .ok() .flatten() .context("Failed to get context server settings")?; if !settings.enabled() { return Ok(()); } let (registry, worktree_store) = this.update(cx, |this, _| { (this.registry.clone(), this.worktree_store.clone()) })?; let configuration = ContextServerConfiguration::from_settings( settings, server.id(), registry, worktree_store, cx, ) .await .context("Failed to create context server configuration")?; this.update(cx, |this, cx| { this.run_server(server, Arc::new(configuration), cx) }) }) .detach_and_log_err(cx); } pub fn stop_server(&mut self, id: &ContextServerId, cx: &mut Context) -> Result<()> { if matches!( self.servers.get(id), Some(ContextServerState::Stopped { .. }) ) { return Ok(()); } let state = self .servers .remove(id) .context("Context server not found")?; let server = state.server(); let configuration = state.configuration(); let mut result = Ok(()); if let ContextServerState::Running { server, .. } = &state { result = server.stop(); } drop(state); self.update_server_state( id.clone(), ContextServerState::Stopped { configuration, server, }, cx, ); result } pub fn restart_server(&mut self, id: &ContextServerId, cx: &mut Context) -> Result<()> { if let Some(state) = self.servers.get(&id) { let configuration = state.configuration(); self.stop_server(&state.server().id(), cx)?; let new_server = self.create_context_server(id.clone(), configuration.clone(), cx); self.run_server(new_server, configuration, cx); } Ok(()) } fn run_server( &mut self, server: Arc, configuration: Arc, cx: &mut Context, ) { let id = server.id(); if matches!( self.servers.get(&id), Some(ContextServerState::Starting { .. } | ContextServerState::Running { .. }) ) { self.stop_server(&id, cx).log_err(); } let task = cx.spawn({ let id = server.id(); let server = server.clone(); let configuration = configuration.clone(); async move |this, cx| { match server.clone().start(&cx).await { Ok(_) => { log::info!("Started {} context server", id); debug_assert!(server.client().is_some()); this.update(cx, |this, cx| { this.update_server_state( id.clone(), ContextServerState::Running { server, configuration, }, cx, ) }) .log_err() } Err(err) => { log::error!("{} context server failed to start: {}", id, err); this.update(cx, |this, cx| { this.update_server_state( id.clone(), ContextServerState::Error { configuration, server, error: err.to_string().into(), }, cx, ) }) .log_err() } }; } }); self.update_server_state( id.clone(), ContextServerState::Starting { configuration, _task: task, server, }, cx, ); } fn remove_server(&mut self, id: &ContextServerId, cx: &mut Context) -> Result<()> { let state = self .servers .remove(id) .context("Context server not found")?; drop(state); cx.emit(Event::ServerStatusChanged { server_id: id.clone(), status: ContextServerStatus::Stopped, }); Ok(()) } fn create_context_server( &self, id: ContextServerId, configuration: Arc, cx: &mut Context, ) -> Arc { let root_path = self .project .read_with(cx, |project, cx| project.active_project_directory(cx)) .ok() .flatten() .or_else(|| { self.worktree_store.read_with(cx, |store, cx| { store.visible_worktrees(cx).fold(None, |acc, item| { if acc.is_none() { item.read(cx).root_dir() } else { acc } }) }) }); if let Some(factory) = self.context_server_factory.as_ref() { factory(id, configuration) } else { Arc::new(ContextServer::stdio( id, configuration.command().clone(), root_path, )) } } fn resolve_context_server_settings<'a>( worktree_store: &'a Entity, cx: &'a App, ) -> &'a HashMap, ContextServerSettings> { let location = worktree_store .read(cx) .visible_worktrees(cx) .next() .map(|worktree| settings::SettingsLocation { worktree_id: worktree.read(cx).id(), path: Path::new(""), }); &ProjectSettings::get(location, cx).context_servers } fn update_server_state( &mut self, id: ContextServerId, state: ContextServerState, cx: &mut Context, ) { let status = ContextServerStatus::from_state(&state); self.servers.insert(id.clone(), state); cx.emit(Event::ServerStatusChanged { server_id: id, status, }); } fn available_context_servers_changed(&mut self, cx: &mut Context) { if self.update_servers_task.is_some() { self.needs_server_update = true; } else { self.needs_server_update = false; self.update_servers_task = Some(cx.spawn(async move |this, cx| { if let Err(err) = Self::maintain_servers(this.clone(), cx).await { log::error!("Error maintaining context servers: {}", err); } this.update(cx, |this, cx| { this.update_servers_task.take(); if this.needs_server_update { this.available_context_servers_changed(cx); } })?; Ok(()) })); } } async fn maintain_servers(this: WeakEntity, cx: &mut AsyncApp) -> Result<()> { let (mut configured_servers, registry, worktree_store) = this.update(cx, |this, _| { ( this.context_server_settings.clone(), this.registry.clone(), this.worktree_store.clone(), ) })?; for (id, _) in registry.read_with(cx, |registry, _| registry.context_server_descriptors())? { configured_servers .entry(id) .or_insert(ContextServerSettings::default_extension()); } let (enabled_servers, disabled_servers): (HashMap<_, _>, HashMap<_, _>) = configured_servers .into_iter() .partition(|(_, settings)| settings.enabled()); let configured_servers = join_all(enabled_servers.into_iter().map(|(id, settings)| { let id = ContextServerId(id); ContextServerConfiguration::from_settings( settings, id.clone(), registry.clone(), worktree_store.clone(), cx, ) .map(|config| (id, config)) })) .await .into_iter() .filter_map(|(id, config)| config.map(|config| (id, config))) .collect::>(); let mut servers_to_start = Vec::new(); let mut servers_to_remove = HashSet::default(); let mut servers_to_stop = HashSet::default(); this.update(cx, |this, cx| { for server_id in this.servers.keys() { // All servers that are not in desired_servers should be removed from the store. // This can happen if the user removed a server from the context server settings. if !configured_servers.contains_key(&server_id) { if disabled_servers.contains_key(&server_id.0) { servers_to_stop.insert(server_id.clone()); } else { servers_to_remove.insert(server_id.clone()); } } } for (id, config) in configured_servers { let state = this.servers.get(&id); let is_stopped = matches!(state, Some(ContextServerState::Stopped { .. })); let existing_config = state.as_ref().map(|state| state.configuration()); if existing_config.as_deref() != Some(&config) || is_stopped { let config = Arc::new(config); let server = this.create_context_server(id.clone(), config.clone(), cx); servers_to_start.push((server, config)); if this.servers.contains_key(&id) { servers_to_stop.insert(id); } } } })?; this.update(cx, |this, cx| { for id in servers_to_stop { this.stop_server(&id, cx)?; } for id in servers_to_remove { this.remove_server(&id, cx)?; } for (server, config) in servers_to_start { this.run_server(server, config, cx); } anyhow::Ok(()) })? } } #[cfg(test)] mod tests { use super::*; use crate::{ FakeFs, Project, context_server_store::registry::ContextServerDescriptor, project_settings::ProjectSettings, }; use context_server::test::create_fake_transport; use gpui::{AppContext, TestAppContext, UpdateGlobal as _}; use serde_json::json; use std::{cell::RefCell, path::PathBuf, rc::Rc}; use util::path; #[gpui::test] async fn test_context_server_status(cx: &mut TestAppContext) { const SERVER_1_ID: &'static str = "mcp-1"; const SERVER_2_ID: &'static str = "mcp-2"; let (_fs, project) = setup_context_server_test( cx, json!({"code.rs": ""}), vec![ (SERVER_1_ID.into(), dummy_server_settings()), (SERVER_2_ID.into(), dummy_server_settings()), ], ) .await; let registry = cx.new(|_| ContextServerDescriptorRegistry::new()); let store = cx.new(|cx| { ContextServerStore::test( registry.clone(), project.read(cx).worktree_store(), project.downgrade(), cx, ) }); let server_1_id = ContextServerId(SERVER_1_ID.into()); let server_2_id = ContextServerId(SERVER_2_ID.into()); let server_1 = Arc::new(ContextServer::new( server_1_id.clone(), Arc::new(create_fake_transport(SERVER_1_ID, cx.executor())), )); let server_2 = Arc::new(ContextServer::new( server_2_id.clone(), Arc::new(create_fake_transport(SERVER_2_ID, cx.executor())), )); store.update(cx, |store, cx| store.start_server(server_1, cx)); cx.run_until_parked(); cx.update(|cx| { assert_eq!( store.read(cx).status_for_server(&server_1_id), Some(ContextServerStatus::Running) ); assert_eq!(store.read(cx).status_for_server(&server_2_id), None); }); store.update(cx, |store, cx| store.start_server(server_2.clone(), cx)); cx.run_until_parked(); cx.update(|cx| { assert_eq!( store.read(cx).status_for_server(&server_1_id), Some(ContextServerStatus::Running) ); assert_eq!( store.read(cx).status_for_server(&server_2_id), Some(ContextServerStatus::Running) ); }); store .update(cx, |store, cx| store.stop_server(&server_2_id, cx)) .unwrap(); cx.update(|cx| { assert_eq!( store.read(cx).status_for_server(&server_1_id), Some(ContextServerStatus::Running) ); assert_eq!( store.read(cx).status_for_server(&server_2_id), Some(ContextServerStatus::Stopped) ); }); } #[gpui::test] async fn test_context_server_status_events(cx: &mut TestAppContext) { const SERVER_1_ID: &'static str = "mcp-1"; const SERVER_2_ID: &'static str = "mcp-2"; let (_fs, project) = setup_context_server_test( cx, json!({"code.rs": ""}), vec![ (SERVER_1_ID.into(), dummy_server_settings()), (SERVER_2_ID.into(), dummy_server_settings()), ], ) .await; let registry = cx.new(|_| ContextServerDescriptorRegistry::new()); let store = cx.new(|cx| { ContextServerStore::test( registry.clone(), project.read(cx).worktree_store(), project.downgrade(), cx, ) }); let server_1_id = ContextServerId(SERVER_1_ID.into()); let server_2_id = ContextServerId(SERVER_2_ID.into()); let server_1 = Arc::new(ContextServer::new( server_1_id.clone(), Arc::new(create_fake_transport(SERVER_1_ID, cx.executor())), )); let server_2 = Arc::new(ContextServer::new( server_2_id.clone(), Arc::new(create_fake_transport(SERVER_2_ID, cx.executor())), )); let _server_events = assert_server_events( &store, vec![ (server_1_id.clone(), ContextServerStatus::Starting), (server_1_id.clone(), ContextServerStatus::Running), (server_2_id.clone(), ContextServerStatus::Starting), (server_2_id.clone(), ContextServerStatus::Running), (server_2_id.clone(), ContextServerStatus::Stopped), ], cx, ); store.update(cx, |store, cx| store.start_server(server_1, cx)); cx.run_until_parked(); store.update(cx, |store, cx| store.start_server(server_2.clone(), cx)); cx.run_until_parked(); store .update(cx, |store, cx| store.stop_server(&server_2_id, cx)) .unwrap(); } #[gpui::test(iterations = 25)] async fn test_context_server_concurrent_starts(cx: &mut TestAppContext) { const SERVER_1_ID: &'static str = "mcp-1"; let (_fs, project) = setup_context_server_test( cx, json!({"code.rs": ""}), vec![(SERVER_1_ID.into(), dummy_server_settings())], ) .await; let registry = cx.new(|_| ContextServerDescriptorRegistry::new()); let store = cx.new(|cx| { ContextServerStore::test( registry.clone(), project.read(cx).worktree_store(), project.downgrade(), cx, ) }); let server_id = ContextServerId(SERVER_1_ID.into()); let server_with_same_id_1 = Arc::new(ContextServer::new( server_id.clone(), Arc::new(create_fake_transport(SERVER_1_ID, cx.executor())), )); let server_with_same_id_2 = Arc::new(ContextServer::new( server_id.clone(), Arc::new(create_fake_transport(SERVER_1_ID, cx.executor())), )); // If we start another server with the same id, we should report that we stopped the previous one let _server_events = assert_server_events( &store, vec![ (server_id.clone(), ContextServerStatus::Starting), (server_id.clone(), ContextServerStatus::Stopped), (server_id.clone(), ContextServerStatus::Starting), (server_id.clone(), ContextServerStatus::Running), ], cx, ); store.update(cx, |store, cx| { store.start_server(server_with_same_id_1.clone(), cx) }); store.update(cx, |store, cx| { store.start_server(server_with_same_id_2.clone(), cx) }); cx.run_until_parked(); cx.update(|cx| { assert_eq!( store.read(cx).status_for_server(&server_id), Some(ContextServerStatus::Running) ); }); } #[gpui::test] async fn test_context_server_maintain_servers_loop(cx: &mut TestAppContext) { const SERVER_1_ID: &'static str = "mcp-1"; const SERVER_2_ID: &'static str = "mcp-2"; let server_1_id = ContextServerId(SERVER_1_ID.into()); let server_2_id = ContextServerId(SERVER_2_ID.into()); let fake_descriptor_1 = Arc::new(FakeContextServerDescriptor::new(SERVER_1_ID)); let (_fs, project) = setup_context_server_test( cx, json!({"code.rs": ""}), vec![( SERVER_1_ID.into(), ContextServerSettings::Extension { enabled: true, settings: json!({ "somevalue": true }), }, )], ) .await; let executor = cx.executor(); let registry = cx.new(|cx| { let mut registry = ContextServerDescriptorRegistry::new(); registry.register_context_server_descriptor(SERVER_1_ID.into(), fake_descriptor_1, cx); registry }); let store = cx.new(|cx| { ContextServerStore::test_maintain_server_loop( Box::new(move |id, _| { Arc::new(ContextServer::new( id.clone(), Arc::new(create_fake_transport(id.0.to_string(), executor.clone())), )) }), registry.clone(), project.read(cx).worktree_store(), project.downgrade(), cx, ) }); // Ensure that mcp-1 starts up { let _server_events = assert_server_events( &store, vec![ (server_1_id.clone(), ContextServerStatus::Starting), (server_1_id.clone(), ContextServerStatus::Running), ], cx, ); cx.run_until_parked(); } // Ensure that mcp-1 is restarted when the configuration was changed { let _server_events = assert_server_events( &store, vec![ (server_1_id.clone(), ContextServerStatus::Stopped), (server_1_id.clone(), ContextServerStatus::Starting), (server_1_id.clone(), ContextServerStatus::Running), ], cx, ); set_context_server_configuration( vec![( server_1_id.0.clone(), ContextServerSettings::Extension { enabled: true, settings: json!({ "somevalue": false }), }, )], cx, ); cx.run_until_parked(); } // Ensure that mcp-1 is not restarted when the configuration was not changed { let _server_events = assert_server_events(&store, vec![], cx); set_context_server_configuration( vec![( server_1_id.0.clone(), ContextServerSettings::Extension { enabled: true, settings: json!({ "somevalue": false }), }, )], cx, ); cx.run_until_parked(); } // Ensure that mcp-2 is started once it is added to the settings { let _server_events = assert_server_events( &store, vec![ (server_2_id.clone(), ContextServerStatus::Starting), (server_2_id.clone(), ContextServerStatus::Running), ], cx, ); set_context_server_configuration( vec![ ( server_1_id.0.clone(), ContextServerSettings::Extension { enabled: true, settings: json!({ "somevalue": false }), }, ), ( server_2_id.0.clone(), ContextServerSettings::Custom { enabled: true, command: ContextServerCommand { path: "somebinary".into(), args: vec!["arg".to_string()], env: None, }, }, ), ], cx, ); cx.run_until_parked(); } // Ensure that mcp-2 is restarted once the args have changed { let _server_events = assert_server_events( &store, vec![ (server_2_id.clone(), ContextServerStatus::Stopped), (server_2_id.clone(), ContextServerStatus::Starting), (server_2_id.clone(), ContextServerStatus::Running), ], cx, ); set_context_server_configuration( vec![ ( server_1_id.0.clone(), ContextServerSettings::Extension { enabled: true, settings: json!({ "somevalue": false }), }, ), ( server_2_id.0.clone(), ContextServerSettings::Custom { enabled: true, command: ContextServerCommand { path: "somebinary".into(), args: vec!["anotherArg".to_string()], env: None, }, }, ), ], cx, ); cx.run_until_parked(); } // Ensure that mcp-2 is removed once it is removed from the settings { let _server_events = assert_server_events( &store, vec![(server_2_id.clone(), ContextServerStatus::Stopped)], cx, ); set_context_server_configuration( vec![( server_1_id.0.clone(), ContextServerSettings::Extension { enabled: true, settings: json!({ "somevalue": false }), }, )], cx, ); cx.run_until_parked(); cx.update(|cx| { assert_eq!(store.read(cx).status_for_server(&server_2_id), None); }); } // Ensure that nothing happens if the settings do not change { let _server_events = assert_server_events(&store, vec![], cx); set_context_server_configuration( vec![( server_1_id.0.clone(), ContextServerSettings::Extension { enabled: true, settings: json!({ "somevalue": false }), }, )], cx, ); cx.run_until_parked(); cx.update(|cx| { assert_eq!( store.read(cx).status_for_server(&server_1_id), Some(ContextServerStatus::Running) ); assert_eq!(store.read(cx).status_for_server(&server_2_id), None); }); } } #[gpui::test] async fn test_context_server_enabled_disabled(cx: &mut TestAppContext) { const SERVER_1_ID: &'static str = "mcp-1"; let server_1_id = ContextServerId(SERVER_1_ID.into()); let (_fs, project) = setup_context_server_test( cx, json!({"code.rs": ""}), vec![( SERVER_1_ID.into(), ContextServerSettings::Custom { enabled: true, command: ContextServerCommand { path: "somebinary".into(), args: vec!["arg".to_string()], env: None, }, }, )], ) .await; let executor = cx.executor(); let registry = cx.new(|_| ContextServerDescriptorRegistry::new()); let store = cx.new(|cx| { ContextServerStore::test_maintain_server_loop( Box::new(move |id, _| { Arc::new(ContextServer::new( id.clone(), Arc::new(create_fake_transport(id.0.to_string(), executor.clone())), )) }), registry.clone(), project.read(cx).worktree_store(), project.downgrade(), cx, ) }); // Ensure that mcp-1 starts up { let _server_events = assert_server_events( &store, vec![ (server_1_id.clone(), ContextServerStatus::Starting), (server_1_id.clone(), ContextServerStatus::Running), ], cx, ); cx.run_until_parked(); } // Ensure that mcp-1 is stopped once it is disabled. { let _server_events = assert_server_events( &store, vec![(server_1_id.clone(), ContextServerStatus::Stopped)], cx, ); set_context_server_configuration( vec![( server_1_id.0.clone(), ContextServerSettings::Custom { enabled: false, command: ContextServerCommand { path: "somebinary".into(), args: vec!["arg".to_string()], env: None, }, }, )], cx, ); cx.run_until_parked(); } // Ensure that mcp-1 is started once it is enabled again. { let _server_events = assert_server_events( &store, vec![ (server_1_id.clone(), ContextServerStatus::Starting), (server_1_id.clone(), ContextServerStatus::Running), ], cx, ); set_context_server_configuration( vec![( server_1_id.0.clone(), ContextServerSettings::Custom { enabled: true, command: ContextServerCommand { path: "somebinary".into(), args: vec!["arg".to_string()], env: None, }, }, )], cx, ); cx.run_until_parked(); } } fn set_context_server_configuration( context_servers: Vec<(Arc, ContextServerSettings)>, cx: &mut TestAppContext, ) { cx.update(|cx| { SettingsStore::update_global(cx, |store, cx| { let mut settings = ProjectSettings::default(); for (id, config) in context_servers { settings.context_servers.insert(id, config); } store .set_user_settings(&serde_json::to_string(&settings).unwrap(), cx) .unwrap(); }) }); } struct ServerEvents { received_event_count: Rc>, expected_event_count: usize, _subscription: Subscription, } impl Drop for ServerEvents { fn drop(&mut self) { let actual_event_count = *self.received_event_count.borrow(); assert_eq!( actual_event_count, self.expected_event_count, " Expected to receive {} context server store events, but received {} events", self.expected_event_count, actual_event_count ); } } fn dummy_server_settings() -> ContextServerSettings { ContextServerSettings::Custom { enabled: true, command: ContextServerCommand { path: "somebinary".into(), args: vec!["arg".to_string()], env: None, }, } } fn assert_server_events( store: &Entity, expected_events: Vec<(ContextServerId, ContextServerStatus)>, cx: &mut TestAppContext, ) -> ServerEvents { cx.update(|cx| { let mut ix = 0; let received_event_count = Rc::new(RefCell::new(0)); let expected_event_count = expected_events.len(); let subscription = cx.subscribe(store, { let received_event_count = received_event_count.clone(); move |_, event, _| match event { Event::ServerStatusChanged { server_id: actual_server_id, status: actual_status, } => { let (expected_server_id, expected_status) = &expected_events[ix]; assert_eq!( actual_server_id, expected_server_id, "Expected different server id at index {}", ix ); assert_eq!( actual_status, expected_status, "Expected different status at index {}", ix ); ix += 1; *received_event_count.borrow_mut() += 1; } } }); ServerEvents { expected_event_count, received_event_count, _subscription: subscription, } }) } async fn setup_context_server_test( cx: &mut TestAppContext, files: serde_json::Value, context_server_configurations: Vec<(Arc, ContextServerSettings)>, ) -> (Arc, Entity) { cx.update(|cx| { let settings_store = SettingsStore::test(cx); cx.set_global(settings_store); Project::init_settings(cx); let mut settings = ProjectSettings::get_global(cx).clone(); for (id, config) in context_server_configurations { settings.context_servers.insert(id, config); } ProjectSettings::override_global(settings, cx); }); let fs = FakeFs::new(cx.executor()); fs.insert_tree(path!("/test"), files).await; let project = Project::test(fs.clone(), [path!("/test").as_ref()], cx).await; (fs, project) } struct FakeContextServerDescriptor { path: PathBuf, } impl FakeContextServerDescriptor { fn new(path: impl Into) -> Self { Self { path: path.into() } } } impl ContextServerDescriptor for FakeContextServerDescriptor { fn command( &self, _worktree_store: Entity, _cx: &AsyncApp, ) -> Task> { Task::ready(Ok(ContextServerCommand { path: self.path.clone(), args: vec!["arg1".to_string(), "arg2".to_string()], env: None, })) } fn configuration( &self, _worktree_store: Entity, _cx: &AsyncApp, ) -> Task>> { Task::ready(Ok(None)) } } }