
- [x] Rewrite worktree git handling - [x] Fix tests - [x] Fix `test_propagate_statuses_for_repos_under_project` - [x] Replace `WorkDirectoryEntry` with `WorkDirectory` in `RepositoryEntry` - [x] Add a worktree event for capturing git status changes - [x] Confirm that the local repositories are correctly updating the new WorkDirectory field - [x] Implement the git statuses query as a join when pulling entries out of worktree - [x] Use this new join to implement the project panel and outline panel. - [x] Synchronize git statuses over the wire for collab and remote dev (use the existing `worktree_repository_statuses` table, adjust as needed) - [x] Only send changed statuses to collab Release Notes: - N/A --------- Co-authored-by: Cole Miller <cole@zed.dev> Co-authored-by: Mikayla <mikayla@zed.com> Co-authored-by: Nathan <nathan@zed.dev>
905 lines
32 KiB
Rust
905 lines
32 KiB
Rust
//! Project-wide storage of the tasks available, capable of updating itself from the sources set.
|
||
|
||
use std::{
|
||
borrow::Cow,
|
||
cmp::{self, Reverse},
|
||
collections::hash_map,
|
||
path::{Path, PathBuf},
|
||
sync::Arc,
|
||
};
|
||
|
||
use anyhow::{Context, Result};
|
||
use collections::{HashMap, HashSet, VecDeque};
|
||
use gpui::{AppContext, Context as _, Model, Task};
|
||
use itertools::Itertools;
|
||
use language::{ContextProvider, File, Language, LanguageToolchainStore, Location};
|
||
use settings::{parse_json_with_comments, SettingsLocation};
|
||
use task::{
|
||
ResolvedTask, TaskContext, TaskId, TaskTemplate, TaskTemplates, TaskVariables, VariableName,
|
||
};
|
||
use text::{Point, ToPoint};
|
||
use util::{post_inc, NumericPrefixWithSuffix, ResultExt as _};
|
||
use worktree::WorktreeId;
|
||
|
||
use crate::worktree_store::WorktreeStore;
|
||
|
||
/// Inventory tracks available tasks for a given project.
|
||
#[derive(Debug, Default)]
|
||
pub struct Inventory {
|
||
last_scheduled_tasks: VecDeque<(TaskSourceKind, ResolvedTask)>,
|
||
templates_from_settings: ParsedTemplates,
|
||
}
|
||
|
||
#[derive(Debug, Default)]
|
||
struct ParsedTemplates {
|
||
global: Vec<TaskTemplate>,
|
||
worktree: HashMap<WorktreeId, HashMap<Arc<Path>, Vec<TaskTemplate>>>,
|
||
}
|
||
|
||
/// Kind of a source the tasks are fetched from, used to display more source information in the UI.
|
||
#[derive(Debug, Clone, PartialEq, Eq, Hash, PartialOrd, Ord)]
|
||
pub enum TaskSourceKind {
|
||
/// bash-like commands spawned by users, not associated with any path
|
||
UserInput,
|
||
/// Tasks from the worktree's .zed/task.json
|
||
Worktree {
|
||
id: WorktreeId,
|
||
directory_in_worktree: PathBuf,
|
||
id_base: Cow<'static, str>,
|
||
},
|
||
/// ~/.config/zed/task.json - like global files with task definitions, applicable to any path
|
||
AbsPath {
|
||
id_base: Cow<'static, str>,
|
||
abs_path: PathBuf,
|
||
},
|
||
/// Languages-specific tasks coming from extensions.
|
||
Language { name: Arc<str> },
|
||
}
|
||
|
||
impl TaskSourceKind {
|
||
pub fn to_id_base(&self) -> String {
|
||
match self {
|
||
TaskSourceKind::UserInput => "oneshot".to_string(),
|
||
TaskSourceKind::AbsPath { id_base, abs_path } => {
|
||
format!("{id_base}_{}", abs_path.display())
|
||
}
|
||
TaskSourceKind::Worktree {
|
||
id,
|
||
id_base,
|
||
directory_in_worktree,
|
||
} => {
|
||
format!("{id_base}_{id}_{}", directory_in_worktree.display())
|
||
}
|
||
TaskSourceKind::Language { name } => format!("language_{name}"),
|
||
}
|
||
}
|
||
}
|
||
|
||
impl Inventory {
|
||
pub fn new(cx: &mut AppContext) -> Model<Self> {
|
||
cx.new_model(|_| Self::default())
|
||
}
|
||
|
||
/// Pulls its task sources relevant to the worktree and the language given,
|
||
/// returns all task templates with their source kinds, worktree tasks first, language tasks second
|
||
/// and global tasks last. No specific order inside source kinds groups.
|
||
pub fn list_tasks(
|
||
&self,
|
||
file: Option<Arc<dyn File>>,
|
||
language: Option<Arc<Language>>,
|
||
worktree: Option<WorktreeId>,
|
||
cx: &AppContext,
|
||
) -> Vec<(TaskSourceKind, TaskTemplate)> {
|
||
let task_source_kind = language.as_ref().map(|language| TaskSourceKind::Language {
|
||
name: language.name().0,
|
||
});
|
||
let global_tasks = self.global_templates_from_settings();
|
||
let language_tasks = language
|
||
.and_then(|language| language.context_provider()?.associated_tasks(file, cx))
|
||
.into_iter()
|
||
.flat_map(|tasks| tasks.0.into_iter())
|
||
.flat_map(|task| Some((task_source_kind.clone()?, task)))
|
||
.chain(global_tasks);
|
||
|
||
self.worktree_templates_from_settings(worktree)
|
||
.chain(language_tasks)
|
||
.collect()
|
||
}
|
||
|
||
/// Pulls its task sources relevant to the worktree and the language given and resolves them with the [`TaskContext`] given.
|
||
/// Joins the new resolutions with the resolved tasks that were used (spawned) before,
|
||
/// orders them so that the most recently used come first, all equally used ones are ordered so that the most specific tasks come first.
|
||
/// Deduplicates the tasks by their labels and context and splits the ordered list into two: used tasks and the rest, newly resolved tasks.
|
||
pub fn used_and_current_resolved_tasks(
|
||
&self,
|
||
worktree: Option<WorktreeId>,
|
||
location: Option<Location>,
|
||
task_context: &TaskContext,
|
||
cx: &AppContext,
|
||
) -> (
|
||
Vec<(TaskSourceKind, ResolvedTask)>,
|
||
Vec<(TaskSourceKind, ResolvedTask)>,
|
||
) {
|
||
let language = location
|
||
.as_ref()
|
||
.and_then(|location| location.buffer.read(cx).language_at(location.range.start));
|
||
let task_source_kind = language.as_ref().map(|language| TaskSourceKind::Language {
|
||
name: language.name().0,
|
||
});
|
||
let file = location
|
||
.as_ref()
|
||
.and_then(|location| location.buffer.read(cx).file().cloned());
|
||
|
||
let mut task_labels_to_ids = HashMap::<String, HashSet<TaskId>>::default();
|
||
let mut lru_score = 0_u32;
|
||
let previously_spawned_tasks = self
|
||
.last_scheduled_tasks
|
||
.iter()
|
||
.rev()
|
||
.filter(|(task_kind, _)| {
|
||
if matches!(task_kind, TaskSourceKind::Language { .. }) {
|
||
Some(task_kind) == task_source_kind.as_ref()
|
||
} else {
|
||
true
|
||
}
|
||
})
|
||
.filter(|(_, resolved_task)| {
|
||
match task_labels_to_ids.entry(resolved_task.resolved_label.clone()) {
|
||
hash_map::Entry::Occupied(mut o) => {
|
||
o.get_mut().insert(resolved_task.id.clone());
|
||
// Neber allow duplicate reused tasks with the same labels
|
||
false
|
||
}
|
||
hash_map::Entry::Vacant(v) => {
|
||
v.insert(HashSet::from_iter(Some(resolved_task.id.clone())));
|
||
true
|
||
}
|
||
}
|
||
})
|
||
.map(|(task_source_kind, resolved_task)| {
|
||
(
|
||
task_source_kind.clone(),
|
||
resolved_task.clone(),
|
||
post_inc(&mut lru_score),
|
||
)
|
||
})
|
||
.sorted_unstable_by(task_lru_comparator)
|
||
.map(|(kind, task, _)| (kind, task))
|
||
.collect::<Vec<_>>();
|
||
|
||
let not_used_score = post_inc(&mut lru_score);
|
||
let global_tasks = self.global_templates_from_settings();
|
||
let language_tasks = language
|
||
.and_then(|language| language.context_provider()?.associated_tasks(file, cx))
|
||
.into_iter()
|
||
.flat_map(|tasks| tasks.0.into_iter())
|
||
.flat_map(|task| Some((task_source_kind.clone()?, task)))
|
||
.chain(global_tasks);
|
||
let worktree_tasks = self
|
||
.worktree_templates_from_settings(worktree)
|
||
.chain(language_tasks);
|
||
|
||
let new_resolved_tasks = worktree_tasks
|
||
.filter_map(|(kind, task)| {
|
||
let id_base = kind.to_id_base();
|
||
Some((
|
||
kind,
|
||
task.resolve_task(&id_base, task_context)?,
|
||
not_used_score,
|
||
))
|
||
})
|
||
.filter(|(_, resolved_task, _)| {
|
||
match task_labels_to_ids.entry(resolved_task.resolved_label.clone()) {
|
||
hash_map::Entry::Occupied(mut o) => {
|
||
// Allow new tasks with the same label, if their context is different
|
||
o.get_mut().insert(resolved_task.id.clone())
|
||
}
|
||
hash_map::Entry::Vacant(v) => {
|
||
v.insert(HashSet::from_iter(Some(resolved_task.id.clone())));
|
||
true
|
||
}
|
||
}
|
||
})
|
||
.sorted_unstable_by(task_lru_comparator)
|
||
.map(|(kind, task, _)| (kind, task))
|
||
.collect::<Vec<_>>();
|
||
|
||
(previously_spawned_tasks, new_resolved_tasks)
|
||
}
|
||
|
||
/// Returns the last scheduled task by task_id if provided.
|
||
/// Otherwise, returns the last scheduled task.
|
||
pub fn last_scheduled_task(
|
||
&self,
|
||
task_id: Option<&TaskId>,
|
||
) -> Option<(TaskSourceKind, ResolvedTask)> {
|
||
if let Some(task_id) = task_id {
|
||
self.last_scheduled_tasks
|
||
.iter()
|
||
.find(|(_, task)| &task.id == task_id)
|
||
.cloned()
|
||
} else {
|
||
self.last_scheduled_tasks.back().cloned()
|
||
}
|
||
}
|
||
|
||
/// Registers task "usage" as being scheduled – to be used for LRU sorting when listing all tasks.
|
||
pub fn task_scheduled(
|
||
&mut self,
|
||
task_source_kind: TaskSourceKind,
|
||
resolved_task: ResolvedTask,
|
||
) {
|
||
self.last_scheduled_tasks
|
||
.push_back((task_source_kind, resolved_task));
|
||
if self.last_scheduled_tasks.len() > 5_000 {
|
||
self.last_scheduled_tasks.pop_front();
|
||
}
|
||
}
|
||
|
||
/// Deletes a resolved task from history, using its id.
|
||
/// A similar may still resurface in `used_and_current_resolved_tasks` when its [`TaskTemplate`] is resolved again.
|
||
pub fn delete_previously_used(&mut self, id: &TaskId) {
|
||
self.last_scheduled_tasks.retain(|(_, task)| &task.id != id);
|
||
}
|
||
|
||
fn global_templates_from_settings(
|
||
&self,
|
||
) -> impl '_ + Iterator<Item = (TaskSourceKind, TaskTemplate)> {
|
||
self.templates_from_settings
|
||
.global
|
||
.clone()
|
||
.into_iter()
|
||
.map(|template| {
|
||
(
|
||
TaskSourceKind::AbsPath {
|
||
id_base: Cow::Borrowed("global tasks.json"),
|
||
abs_path: paths::tasks_file().clone(),
|
||
},
|
||
template,
|
||
)
|
||
})
|
||
}
|
||
|
||
fn worktree_templates_from_settings(
|
||
&self,
|
||
worktree: Option<WorktreeId>,
|
||
) -> impl '_ + Iterator<Item = (TaskSourceKind, TaskTemplate)> {
|
||
worktree.into_iter().flat_map(|worktree| {
|
||
self.templates_from_settings
|
||
.worktree
|
||
.get(&worktree)
|
||
.into_iter()
|
||
.flatten()
|
||
.flat_map(|(directory, templates)| {
|
||
templates.iter().map(move |template| (directory, template))
|
||
})
|
||
.map(move |(directory, template)| {
|
||
(
|
||
TaskSourceKind::Worktree {
|
||
id: worktree,
|
||
directory_in_worktree: directory.to_path_buf(),
|
||
id_base: Cow::Owned(format!(
|
||
"local worktree tasks from directory {directory:?}"
|
||
)),
|
||
},
|
||
template.clone(),
|
||
)
|
||
})
|
||
})
|
||
}
|
||
|
||
/// Updates in-memory task metadata from the JSON string given.
|
||
/// Will fail if the JSON is not a valid array of objects, but will continue if any object will not parse into a [`TaskTemplate`].
|
||
///
|
||
/// Global tasks are updated for no worktree provided, otherwise the worktree metadata for a given path will be updated.
|
||
pub(crate) fn update_file_based_tasks(
|
||
&mut self,
|
||
location: Option<SettingsLocation<'_>>,
|
||
raw_tasks_json: Option<&str>,
|
||
) -> anyhow::Result<()> {
|
||
let raw_tasks =
|
||
parse_json_with_comments::<Vec<serde_json::Value>>(raw_tasks_json.unwrap_or("[]"))
|
||
.context("parsing tasks file content as a JSON array")?;
|
||
let new_templates = raw_tasks.into_iter().filter_map(|raw_template| {
|
||
serde_json::from_value::<TaskTemplate>(raw_template).log_err()
|
||
});
|
||
|
||
let parsed_templates = &mut self.templates_from_settings;
|
||
match location {
|
||
Some(location) => {
|
||
let new_templates = new_templates.collect::<Vec<_>>();
|
||
if new_templates.is_empty() {
|
||
if let Some(worktree_tasks) =
|
||
parsed_templates.worktree.get_mut(&location.worktree_id)
|
||
{
|
||
worktree_tasks.remove(location.path);
|
||
}
|
||
} else {
|
||
parsed_templates
|
||
.worktree
|
||
.entry(location.worktree_id)
|
||
.or_default()
|
||
.insert(Arc::from(location.path), new_templates);
|
||
}
|
||
}
|
||
None => parsed_templates.global = new_templates.collect(),
|
||
}
|
||
Ok(())
|
||
}
|
||
}
|
||
|
||
fn task_lru_comparator(
|
||
(kind_a, task_a, lru_score_a): &(TaskSourceKind, ResolvedTask, u32),
|
||
(kind_b, task_b, lru_score_b): &(TaskSourceKind, ResolvedTask, u32),
|
||
) -> cmp::Ordering {
|
||
lru_score_a
|
||
// First, display recently used templates above all.
|
||
.cmp(lru_score_b)
|
||
// Then, ensure more specific sources are displayed first.
|
||
.then(task_source_kind_preference(kind_a).cmp(&task_source_kind_preference(kind_b)))
|
||
// After that, display first more specific tasks, using more template variables.
|
||
// Bonus points for tasks with symbol variables.
|
||
.then(task_variables_preference(task_a).cmp(&task_variables_preference(task_b)))
|
||
// Finally, sort by the resolved label, but a bit more specifically, to avoid mixing letters and digits.
|
||
.then({
|
||
NumericPrefixWithSuffix::from_numeric_prefixed_str(&task_a.resolved_label)
|
||
.cmp(&NumericPrefixWithSuffix::from_numeric_prefixed_str(
|
||
&task_b.resolved_label,
|
||
))
|
||
.then(task_a.resolved_label.cmp(&task_b.resolved_label))
|
||
.then(kind_a.cmp(kind_b))
|
||
})
|
||
}
|
||
|
||
fn task_source_kind_preference(kind: &TaskSourceKind) -> u32 {
|
||
match kind {
|
||
TaskSourceKind::Language { .. } => 1,
|
||
TaskSourceKind::UserInput => 2,
|
||
TaskSourceKind::Worktree { .. } => 3,
|
||
TaskSourceKind::AbsPath { .. } => 4,
|
||
}
|
||
}
|
||
|
||
fn task_variables_preference(task: &ResolvedTask) -> Reverse<usize> {
|
||
let task_variables = task.substituted_variables();
|
||
Reverse(if task_variables.contains(&VariableName::Symbol) {
|
||
task_variables.len() + 1
|
||
} else {
|
||
task_variables.len()
|
||
})
|
||
}
|
||
|
||
#[cfg(test)]
|
||
mod test_inventory {
|
||
use gpui::{Model, TestAppContext};
|
||
use itertools::Itertools;
|
||
use task::TaskContext;
|
||
use worktree::WorktreeId;
|
||
|
||
use crate::Inventory;
|
||
|
||
use super::TaskSourceKind;
|
||
|
||
pub(super) fn task_template_names(
|
||
inventory: &Model<Inventory>,
|
||
worktree: Option<WorktreeId>,
|
||
cx: &mut TestAppContext,
|
||
) -> Vec<String> {
|
||
inventory.update(cx, |inventory, cx| {
|
||
inventory
|
||
.list_tasks(None, None, worktree, cx)
|
||
.into_iter()
|
||
.map(|(_, task)| task.label)
|
||
.sorted()
|
||
.collect()
|
||
})
|
||
}
|
||
|
||
pub(super) fn register_task_used(
|
||
inventory: &Model<Inventory>,
|
||
task_name: &str,
|
||
cx: &mut TestAppContext,
|
||
) {
|
||
inventory.update(cx, |inventory, cx| {
|
||
let (task_source_kind, task) = inventory
|
||
.list_tasks(None, None, None, cx)
|
||
.into_iter()
|
||
.find(|(_, task)| task.label == task_name)
|
||
.unwrap_or_else(|| panic!("Failed to find task with name {task_name}"));
|
||
let id_base = task_source_kind.to_id_base();
|
||
inventory.task_scheduled(
|
||
task_source_kind.clone(),
|
||
task.resolve_task(&id_base, &TaskContext::default())
|
||
.unwrap_or_else(|| panic!("Failed to resolve task with name {task_name}")),
|
||
);
|
||
});
|
||
}
|
||
|
||
pub(super) async fn list_tasks(
|
||
inventory: &Model<Inventory>,
|
||
worktree: Option<WorktreeId>,
|
||
cx: &mut TestAppContext,
|
||
) -> Vec<(TaskSourceKind, String)> {
|
||
inventory.update(cx, |inventory, cx| {
|
||
let task_context = &TaskContext::default();
|
||
inventory
|
||
.list_tasks(None, None, worktree, cx)
|
||
.into_iter()
|
||
.filter_map(|(source_kind, task)| {
|
||
let id_base = source_kind.to_id_base();
|
||
Some((source_kind, task.resolve_task(&id_base, task_context)?))
|
||
})
|
||
.map(|(source_kind, resolved_task)| (source_kind, resolved_task.resolved_label))
|
||
.collect()
|
||
})
|
||
}
|
||
}
|
||
|
||
/// A context provided that tries to provide values for all non-custom [`VariableName`] variants for a currently opened file.
|
||
/// Applied as a base for every custom [`ContextProvider`] unless explicitly oped out.
|
||
pub struct BasicContextProvider {
|
||
worktree_store: Model<WorktreeStore>,
|
||
}
|
||
|
||
impl BasicContextProvider {
|
||
pub fn new(worktree_store: Model<WorktreeStore>) -> Self {
|
||
Self { worktree_store }
|
||
}
|
||
}
|
||
impl ContextProvider for BasicContextProvider {
|
||
fn build_context(
|
||
&self,
|
||
_: &TaskVariables,
|
||
location: &Location,
|
||
_: Option<HashMap<String, String>>,
|
||
_: Arc<dyn LanguageToolchainStore>,
|
||
cx: &mut AppContext,
|
||
) -> Task<Result<TaskVariables>> {
|
||
let buffer = location.buffer.read(cx);
|
||
let buffer_snapshot = buffer.snapshot();
|
||
let symbols = buffer_snapshot.symbols_containing(location.range.start, None);
|
||
let symbol = symbols.unwrap_or_default().last().map(|symbol| {
|
||
let range = symbol
|
||
.name_ranges
|
||
.last()
|
||
.cloned()
|
||
.unwrap_or(0..symbol.text.len());
|
||
symbol.text[range].to_string()
|
||
});
|
||
|
||
let current_file = buffer
|
||
.file()
|
||
.and_then(|file| file.as_local())
|
||
.map(|file| file.abs_path(cx).to_string_lossy().to_string());
|
||
let Point { row, column } = location.range.start.to_point(&buffer_snapshot);
|
||
let row = row + 1;
|
||
let column = column + 1;
|
||
let selected_text = buffer
|
||
.chars_for_range(location.range.clone())
|
||
.collect::<String>();
|
||
|
||
let mut task_variables = TaskVariables::from_iter([
|
||
(VariableName::Row, row.to_string()),
|
||
(VariableName::Column, column.to_string()),
|
||
]);
|
||
|
||
if let Some(symbol) = symbol {
|
||
task_variables.insert(VariableName::Symbol, symbol);
|
||
}
|
||
if !selected_text.trim().is_empty() {
|
||
task_variables.insert(VariableName::SelectedText, selected_text);
|
||
}
|
||
let worktree_abs_path =
|
||
buffer
|
||
.file()
|
||
.map(|file| file.worktree_id(cx))
|
||
.and_then(|worktree_id| {
|
||
self.worktree_store
|
||
.read(cx)
|
||
.worktree_for_id(worktree_id, cx)
|
||
.map(|worktree| worktree.read(cx).abs_path())
|
||
});
|
||
if let Some(worktree_path) = worktree_abs_path {
|
||
task_variables.insert(
|
||
VariableName::WorktreeRoot,
|
||
worktree_path.to_string_lossy().to_string(),
|
||
);
|
||
if let Some(full_path) = current_file.as_ref() {
|
||
let relative_path = pathdiff::diff_paths(full_path, worktree_path);
|
||
if let Some(relative_path) = relative_path {
|
||
task_variables.insert(
|
||
VariableName::RelativeFile,
|
||
relative_path.to_string_lossy().into_owned(),
|
||
);
|
||
}
|
||
}
|
||
}
|
||
|
||
if let Some(path_as_string) = current_file {
|
||
let path = Path::new(&path_as_string);
|
||
if let Some(filename) = path.file_name().and_then(|f| f.to_str()) {
|
||
task_variables.insert(VariableName::Filename, String::from(filename));
|
||
}
|
||
|
||
if let Some(stem) = path.file_stem().and_then(|s| s.to_str()) {
|
||
task_variables.insert(VariableName::Stem, stem.into());
|
||
}
|
||
|
||
if let Some(dirname) = path.parent().and_then(|s| s.to_str()) {
|
||
task_variables.insert(VariableName::Dirname, dirname.into());
|
||
}
|
||
|
||
task_variables.insert(VariableName::File, path_as_string);
|
||
}
|
||
|
||
Task::ready(Ok(task_variables))
|
||
}
|
||
}
|
||
|
||
/// A ContextProvider that doesn't provide any task variables on it's own, though it has some associated tasks.
|
||
pub struct ContextProviderWithTasks {
|
||
templates: TaskTemplates,
|
||
}
|
||
|
||
impl ContextProviderWithTasks {
|
||
pub fn new(definitions: TaskTemplates) -> Self {
|
||
Self {
|
||
templates: definitions,
|
||
}
|
||
}
|
||
}
|
||
|
||
impl ContextProvider for ContextProviderWithTasks {
|
||
fn associated_tasks(
|
||
&self,
|
||
_: Option<Arc<dyn language::File>>,
|
||
_: &AppContext,
|
||
) -> Option<TaskTemplates> {
|
||
Some(self.templates.clone())
|
||
}
|
||
}
|
||
|
||
#[cfg(test)]
|
||
mod tests {
|
||
use gpui::TestAppContext;
|
||
use pretty_assertions::assert_eq;
|
||
use serde_json::json;
|
||
|
||
use crate::task_store::TaskStore;
|
||
|
||
use super::test_inventory::*;
|
||
use super::*;
|
||
|
||
#[gpui::test]
|
||
async fn test_task_list_sorting(cx: &mut TestAppContext) {
|
||
init_test(cx);
|
||
let inventory = cx.update(Inventory::new);
|
||
let initial_tasks = resolved_task_names(&inventory, None, cx).await;
|
||
assert!(
|
||
initial_tasks.is_empty(),
|
||
"No tasks expected for empty inventory, but got {initial_tasks:?}"
|
||
);
|
||
let initial_tasks = task_template_names(&inventory, None, cx);
|
||
assert!(
|
||
initial_tasks.is_empty(),
|
||
"No tasks expected for empty inventory, but got {initial_tasks:?}"
|
||
);
|
||
cx.run_until_parked();
|
||
let expected_initial_state = [
|
||
"1_a_task".to_string(),
|
||
"1_task".to_string(),
|
||
"2_task".to_string(),
|
||
"3_task".to_string(),
|
||
];
|
||
|
||
inventory.update(cx, |inventory, _| {
|
||
inventory
|
||
.update_file_based_tasks(
|
||
None,
|
||
Some(&mock_tasks_from_names(
|
||
expected_initial_state.iter().map(|name| name.as_str()),
|
||
)),
|
||
)
|
||
.unwrap();
|
||
});
|
||
assert_eq!(
|
||
task_template_names(&inventory, None, cx),
|
||
&expected_initial_state,
|
||
);
|
||
assert_eq!(
|
||
resolved_task_names(&inventory, None, cx).await,
|
||
&expected_initial_state,
|
||
"Tasks with equal amount of usages should be sorted alphanumerically"
|
||
);
|
||
|
||
register_task_used(&inventory, "2_task", cx);
|
||
assert_eq!(
|
||
task_template_names(&inventory, None, cx),
|
||
&expected_initial_state,
|
||
);
|
||
assert_eq!(
|
||
resolved_task_names(&inventory, None, cx).await,
|
||
vec![
|
||
"2_task".to_string(),
|
||
"1_a_task".to_string(),
|
||
"1_task".to_string(),
|
||
"3_task".to_string()
|
||
],
|
||
);
|
||
|
||
register_task_used(&inventory, "1_task", cx);
|
||
register_task_used(&inventory, "1_task", cx);
|
||
register_task_used(&inventory, "1_task", cx);
|
||
register_task_used(&inventory, "3_task", cx);
|
||
assert_eq!(
|
||
task_template_names(&inventory, None, cx),
|
||
&expected_initial_state,
|
||
);
|
||
assert_eq!(
|
||
resolved_task_names(&inventory, None, cx).await,
|
||
vec![
|
||
"3_task".to_string(),
|
||
"1_task".to_string(),
|
||
"2_task".to_string(),
|
||
"1_a_task".to_string(),
|
||
],
|
||
);
|
||
|
||
inventory.update(cx, |inventory, _| {
|
||
inventory
|
||
.update_file_based_tasks(
|
||
None,
|
||
Some(&mock_tasks_from_names(
|
||
["10_hello", "11_hello"]
|
||
.into_iter()
|
||
.chain(expected_initial_state.iter().map(|name| name.as_str())),
|
||
)),
|
||
)
|
||
.unwrap();
|
||
});
|
||
cx.run_until_parked();
|
||
let expected_updated_state = [
|
||
"10_hello".to_string(),
|
||
"11_hello".to_string(),
|
||
"1_a_task".to_string(),
|
||
"1_task".to_string(),
|
||
"2_task".to_string(),
|
||
"3_task".to_string(),
|
||
];
|
||
assert_eq!(
|
||
task_template_names(&inventory, None, cx),
|
||
&expected_updated_state,
|
||
);
|
||
assert_eq!(
|
||
resolved_task_names(&inventory, None, cx).await,
|
||
vec![
|
||
"3_task".to_string(),
|
||
"1_task".to_string(),
|
||
"2_task".to_string(),
|
||
"1_a_task".to_string(),
|
||
"10_hello".to_string(),
|
||
"11_hello".to_string(),
|
||
],
|
||
);
|
||
|
||
register_task_used(&inventory, "11_hello", cx);
|
||
assert_eq!(
|
||
task_template_names(&inventory, None, cx),
|
||
&expected_updated_state,
|
||
);
|
||
assert_eq!(
|
||
resolved_task_names(&inventory, None, cx).await,
|
||
vec![
|
||
"11_hello".to_string(),
|
||
"3_task".to_string(),
|
||
"1_task".to_string(),
|
||
"2_task".to_string(),
|
||
"1_a_task".to_string(),
|
||
"10_hello".to_string(),
|
||
],
|
||
);
|
||
}
|
||
|
||
#[gpui::test]
|
||
async fn test_inventory_static_task_filters(cx: &mut TestAppContext) {
|
||
init_test(cx);
|
||
let inventory = cx.update(Inventory::new);
|
||
let common_name = "common_task_name";
|
||
let worktree_1 = WorktreeId::from_usize(1);
|
||
let worktree_2 = WorktreeId::from_usize(2);
|
||
|
||
cx.run_until_parked();
|
||
let worktree_independent_tasks = vec![
|
||
(
|
||
TaskSourceKind::AbsPath {
|
||
id_base: "global tasks.json".into(),
|
||
abs_path: paths::tasks_file().clone(),
|
||
},
|
||
common_name.to_string(),
|
||
),
|
||
(
|
||
TaskSourceKind::AbsPath {
|
||
id_base: "global tasks.json".into(),
|
||
abs_path: paths::tasks_file().clone(),
|
||
},
|
||
"static_source_1".to_string(),
|
||
),
|
||
(
|
||
TaskSourceKind::AbsPath {
|
||
id_base: "global tasks.json".into(),
|
||
abs_path: paths::tasks_file().clone(),
|
||
},
|
||
"static_source_2".to_string(),
|
||
),
|
||
];
|
||
let worktree_1_tasks = [
|
||
(
|
||
TaskSourceKind::Worktree {
|
||
id: worktree_1,
|
||
directory_in_worktree: PathBuf::from(".zed"),
|
||
id_base: "local worktree tasks from directory \".zed\"".into(),
|
||
},
|
||
common_name.to_string(),
|
||
),
|
||
(
|
||
TaskSourceKind::Worktree {
|
||
id: worktree_1,
|
||
directory_in_worktree: PathBuf::from(".zed"),
|
||
id_base: "local worktree tasks from directory \".zed\"".into(),
|
||
},
|
||
"worktree_1".to_string(),
|
||
),
|
||
];
|
||
let worktree_2_tasks = [
|
||
(
|
||
TaskSourceKind::Worktree {
|
||
id: worktree_2,
|
||
directory_in_worktree: PathBuf::from(".zed"),
|
||
id_base: "local worktree tasks from directory \".zed\"".into(),
|
||
},
|
||
common_name.to_string(),
|
||
),
|
||
(
|
||
TaskSourceKind::Worktree {
|
||
id: worktree_2,
|
||
directory_in_worktree: PathBuf::from(".zed"),
|
||
id_base: "local worktree tasks from directory \".zed\"".into(),
|
||
},
|
||
"worktree_2".to_string(),
|
||
),
|
||
];
|
||
|
||
inventory.update(cx, |inventory, _| {
|
||
inventory
|
||
.update_file_based_tasks(
|
||
None,
|
||
Some(&mock_tasks_from_names(
|
||
worktree_independent_tasks
|
||
.iter()
|
||
.map(|(_, name)| name.as_str()),
|
||
)),
|
||
)
|
||
.unwrap();
|
||
inventory
|
||
.update_file_based_tasks(
|
||
Some(SettingsLocation {
|
||
worktree_id: worktree_1,
|
||
path: Path::new(".zed"),
|
||
}),
|
||
Some(&mock_tasks_from_names(
|
||
worktree_1_tasks.iter().map(|(_, name)| name.as_str()),
|
||
)),
|
||
)
|
||
.unwrap();
|
||
inventory
|
||
.update_file_based_tasks(
|
||
Some(SettingsLocation {
|
||
worktree_id: worktree_2,
|
||
path: Path::new(".zed"),
|
||
}),
|
||
Some(&mock_tasks_from_names(
|
||
worktree_2_tasks.iter().map(|(_, name)| name.as_str()),
|
||
)),
|
||
)
|
||
.unwrap();
|
||
});
|
||
|
||
assert_eq!(
|
||
list_tasks_sorted_by_last_used(&inventory, None, cx).await,
|
||
worktree_independent_tasks,
|
||
"Without a worktree, only worktree-independent tasks should be listed"
|
||
);
|
||
assert_eq!(
|
||
list_tasks_sorted_by_last_used(&inventory, Some(worktree_1), cx).await,
|
||
worktree_1_tasks
|
||
.iter()
|
||
.chain(worktree_independent_tasks.iter())
|
||
.cloned()
|
||
.sorted_by_key(|(kind, label)| (task_source_kind_preference(kind), label.clone()))
|
||
.collect::<Vec<_>>(),
|
||
);
|
||
assert_eq!(
|
||
list_tasks_sorted_by_last_used(&inventory, Some(worktree_2), cx).await,
|
||
worktree_2_tasks
|
||
.iter()
|
||
.chain(worktree_independent_tasks.iter())
|
||
.cloned()
|
||
.sorted_by_key(|(kind, label)| (task_source_kind_preference(kind), label.clone()))
|
||
.collect::<Vec<_>>(),
|
||
);
|
||
|
||
assert_eq!(
|
||
list_tasks(&inventory, None, cx).await,
|
||
worktree_independent_tasks,
|
||
"Without a worktree, only worktree-independent tasks should be listed"
|
||
);
|
||
assert_eq!(
|
||
list_tasks(&inventory, Some(worktree_1), cx).await,
|
||
worktree_1_tasks
|
||
.iter()
|
||
.chain(worktree_independent_tasks.iter())
|
||
.cloned()
|
||
.collect::<Vec<_>>(),
|
||
);
|
||
assert_eq!(
|
||
list_tasks(&inventory, Some(worktree_2), cx).await,
|
||
worktree_2_tasks
|
||
.iter()
|
||
.chain(worktree_independent_tasks.iter())
|
||
.cloned()
|
||
.collect::<Vec<_>>(),
|
||
);
|
||
}
|
||
|
||
fn init_test(_cx: &mut TestAppContext) {
|
||
if std::env::var("RUST_LOG").is_ok() {
|
||
env_logger::try_init().ok();
|
||
}
|
||
TaskStore::init(None);
|
||
}
|
||
|
||
async fn resolved_task_names(
|
||
inventory: &Model<Inventory>,
|
||
worktree: Option<WorktreeId>,
|
||
cx: &mut TestAppContext,
|
||
) -> Vec<String> {
|
||
let (used, current) = inventory.update(cx, |inventory, cx| {
|
||
inventory.used_and_current_resolved_tasks(worktree, None, &TaskContext::default(), cx)
|
||
});
|
||
used.into_iter()
|
||
.chain(current)
|
||
.map(|(_, task)| task.original_task().label.clone())
|
||
.collect()
|
||
}
|
||
|
||
fn mock_tasks_from_names<'a>(task_names: impl Iterator<Item = &'a str> + 'a) -> String {
|
||
serde_json::to_string(&serde_json::Value::Array(
|
||
task_names
|
||
.map(|task_name| {
|
||
json!({
|
||
"label": task_name,
|
||
"command": "echo",
|
||
"args": vec![task_name],
|
||
})
|
||
})
|
||
.collect::<Vec<_>>(),
|
||
))
|
||
.unwrap()
|
||
}
|
||
|
||
async fn list_tasks_sorted_by_last_used(
|
||
inventory: &Model<Inventory>,
|
||
worktree: Option<WorktreeId>,
|
||
cx: &mut TestAppContext,
|
||
) -> Vec<(TaskSourceKind, String)> {
|
||
let (used, current) = inventory.update(cx, |inventory, cx| {
|
||
inventory.used_and_current_resolved_tasks(worktree, None, &TaskContext::default(), cx)
|
||
});
|
||
let mut all = used;
|
||
all.extend(current);
|
||
all.into_iter()
|
||
.map(|(source_kind, task)| (source_kind, task.resolved_label))
|
||
.sorted_by_key(|(kind, label)| (task_source_kind_preference(kind), label.clone()))
|
||
.collect()
|
||
}
|
||
}
|