Compare commits

...
Sign in to create a new pull request.

6 commits

9 changed files with 231 additions and 109 deletions

View file

@ -156,6 +156,9 @@ pub struct PredictEditsBody {
#[derive(Debug, Clone, Serialize, Deserialize)]
pub struct PredictEditsGitInfo {
/// Repo-relative path to the file that contains the input excerpt.
#[serde(skip_serializing_if = "Option::is_none", default)]
pub input_file_path: Option<String>,
/// SHA of git HEAD commit at time of prediction.
#[serde(skip_serializing_if = "Option::is_none", default)]
pub head_sha: Option<String>,
@ -165,6 +168,17 @@ pub struct PredictEditsGitInfo {
/// URL of the remote called `upstream`.
#[serde(skip_serializing_if = "Option::is_none", default)]
pub remote_upstream_url: Option<String>,
/// Recently active files that may be within this repository.
#[serde(skip_serializing_if = "Option::is_none", default)]
pub recent_files: Option<Vec<PredictEditsRecentFile>>,
}
#[derive(Debug, Clone, Serialize, Deserialize)]
pub struct PredictEditsRecentFile {
/// Path to a file within the repository.
pub path: String,
/// Milliseconds between the editor for this file being active and the request time.
pub active_to_now_ms: u32,
}
#[derive(Debug, Clone, Serialize, Deserialize)]

View file

@ -3072,12 +3072,7 @@ async fn test_git_status_sync(
.collect::<Vec<_>>();
assert_eq!(repos.len(), 1);
let repo = repos.into_iter().next().unwrap();
assert_eq!(
repo.read(cx)
.status_for_path(&file.into())
.map(|entry| entry.status),
status
);
assert_eq!(repo.read(cx).status_for_path(&file.into()), status);
}
project_local.read_with(cx_a, |project, cx| {

View file

@ -685,7 +685,7 @@ impl Item for Editor {
.git_store()
.read(cx)
.repository_and_path_for_buffer_id(buffer_id, cx)?;
let status = repo.read(cx).status_for_path(&repo_path)?.status;
let status = repo.read(cx).status_for_path(&repo_path)?;
Some(entry_git_aware_label_color(
status.summary(),

View file

@ -2693,9 +2693,7 @@ impl OutlinePanel {
let status = git_store
.read(cx)
.repository_and_path_for_buffer_id(buffer_id, cx)
.and_then(|(repo, path)| {
Some(repo.read(cx).status_for_path(&path)?.status)
});
.and_then(|(repo, path)| Some(repo.read(cx).status_for_path(&path)?));
buffer_excerpts
.entry(buffer_id)
.or_insert_with(|| {

View file

@ -809,7 +809,7 @@ impl GitStore {
cx: &App,
) -> Option<FileStatus> {
let (repo, repo_path) = self.repository_and_path_for_project_path(project_path, cx)?;
Some(repo.read(cx).status_for_path(&repo_path)?.status)
Some(repo.read(cx).status_for_path(&repo_path)?)
}
pub fn checkpoint(&self, cx: &mut App) -> Task<Result<GitStoreCheckpoint>> {
@ -1391,8 +1391,7 @@ impl GitStore {
pub fn status_for_buffer_id(&self, buffer_id: BufferId, cx: &App) -> Option<FileStatus> {
let (repo, path) = self.repository_and_path_for_buffer_id(buffer_id, cx)?;
let status = repo.read(cx).snapshot.status_for_path(&path)?;
Some(status.status)
repo.read(cx).snapshot.status_for_path(&path)
}
pub fn repository_and_path_for_buffer_id(
@ -2844,10 +2843,10 @@ impl RepositorySnapshot {
self.statuses_by_path.summary().item_summary
}
pub fn status_for_path(&self, path: &RepoPath) -> Option<StatusEntry> {
pub fn status_for_path(&self, path: &RepoPath) -> Option<FileStatus> {
self.statuses_by_path
.get(&PathKey(path.0.clone()), &())
.cloned()
.map(|entry| entry.status.clone())
}
pub fn abs_path_to_repo_path(&self, abs_path: &Path) -> Option<RepoPath> {
@ -2874,7 +2873,7 @@ impl RepositorySnapshot {
self.merge.conflicted_paths.contains(repo_path);
let has_conflict_currently = self
.status_for_path(repo_path)
.is_some_and(|entry| entry.status.is_conflicted());
.is_some_and(|status| status.is_conflicted());
had_conflict_on_last_merge_head_change || has_conflict_currently
}

View file

@ -8246,7 +8246,7 @@ async fn test_repository_subfolder_git_status(
assert_eq!(repository.status_for_path(&C_TXT.into()), None);
assert_eq!(
repository.status_for_path(&E_TXT.into()).unwrap().status,
repository.status_for_path(&E_TXT.into()).unwrap(),
FileStatus::Untracked
);
});
@ -8459,15 +8459,11 @@ async fn test_rename_work_directory(cx: &mut gpui::TestAppContext) {
root_path.join("projects/project1").as_path()
);
assert_eq!(
repository
.status_for_path(&"a".into())
.map(|entry| entry.status),
repository.status_for_path(&"a".into()),
Some(StatusCode::Modified.worktree()),
);
assert_eq!(
repository
.status_for_path(&"b".into())
.map(|entry| entry.status),
repository.status_for_path(&"b".into()),
Some(FileStatus::Untracked),
);
});
@ -8485,11 +8481,11 @@ async fn test_rename_work_directory(cx: &mut gpui::TestAppContext) {
root_path.join("projects/project2").as_path()
);
assert_eq!(
repository.status_for_path(&"a".into()).unwrap().status,
repository.status_for_path(&"a".into()).unwrap(),
StatusCode::Modified.worktree(),
);
assert_eq!(
repository.status_for_path(&"b".into()).unwrap().status,
repository.status_for_path(&"b".into()).unwrap(),
FileStatus::Untracked,
);
});
@ -8562,11 +8558,11 @@ async fn test_file_status(cx: &mut gpui::TestAppContext) {
);
assert_eq!(
repository.status_for_path(&B_TXT.into()).unwrap().status,
repository.status_for_path(&B_TXT.into()).unwrap(),
FileStatus::Untracked,
);
assert_eq!(
repository.status_for_path(&F_TXT.into()).unwrap().status,
repository.status_for_path(&F_TXT.into()).unwrap(),
FileStatus::Untracked,
);
});
@ -8582,7 +8578,7 @@ async fn test_file_status(cx: &mut gpui::TestAppContext) {
// The worktree detects that the file's git status has changed.
repository.read_with(cx, |repository, _| {
assert_eq!(
repository.status_for_path(&A_TXT.into()).unwrap().status,
repository.status_for_path(&A_TXT.into()).unwrap(),
StatusCode::Modified.worktree(),
);
});
@ -8600,7 +8596,7 @@ async fn test_file_status(cx: &mut gpui::TestAppContext) {
// The worktree detects that the files' git status have changed.
repository.read_with(cx, |repository, _cx| {
assert_eq!(
repository.status_for_path(&F_TXT.into()).unwrap().status,
repository.status_for_path(&F_TXT.into()).unwrap(),
FileStatus::Untracked,
);
assert_eq!(repository.status_for_path(&B_TXT.into()), None);
@ -8623,11 +8619,11 @@ async fn test_file_status(cx: &mut gpui::TestAppContext) {
repository.read_with(cx, |repository, _cx| {
assert_eq!(repository.status_for_path(&A_TXT.into()), None);
assert_eq!(
repository.status_for_path(&B_TXT.into()).unwrap().status,
repository.status_for_path(&B_TXT.into()).unwrap(),
FileStatus::Untracked,
);
assert_eq!(
repository.status_for_path(&E_TXT.into()).unwrap().status,
repository.status_for_path(&E_TXT.into()).unwrap(),
StatusCode::Modified.worktree(),
);
});
@ -8666,8 +8662,7 @@ async fn test_file_status(cx: &mut gpui::TestAppContext) {
assert_eq!(
repository
.status_for_path(&Path::new(renamed_dir_name).join(RENAMED_FILE).into())
.unwrap()
.status,
.unwrap(),
FileStatus::Untracked,
);
});
@ -8690,8 +8685,7 @@ async fn test_file_status(cx: &mut gpui::TestAppContext) {
assert_eq!(
repository
.status_for_path(&Path::new(renamed_dir_name).join(RENAMED_FILE).into())
.unwrap()
.status,
.unwrap(),
FileStatus::Untracked,
);
});
@ -9000,7 +8994,7 @@ async fn test_git_worktrees_and_submodules(cx: &mut gpui::TestAppContext) {
barrier.await.unwrap();
worktree_repo.update(cx, |repo, _| {
pretty_assertions::assert_eq!(
repo.status_for_path(&"src/b.txt".into()).unwrap().status,
repo.status_for_path(&"src/b.txt".into()).unwrap(),
StatusCode::Modified.worktree(),
);
});
@ -9039,7 +9033,7 @@ async fn test_git_worktrees_and_submodules(cx: &mut gpui::TestAppContext) {
barrier.await.unwrap();
submodule_repo.update(cx, |repo, _| {
pretty_assertions::assert_eq!(
repo.status_for_path(&"c.txt".into()).unwrap().status,
repo.status_for_path(&"c.txt".into()).unwrap(),
StatusCode::Modified.worktree(),
);
});
@ -9300,9 +9294,7 @@ fn assert_entry_git_state(
let entry = tree
.entry_for_path(path)
.unwrap_or_else(|| panic!("entry {path} not found"));
let status = repository
.status_for_path(&path.into())
.map(|entry| entry.status);
let status = repository.status_for_path(&path.into()).map(|entry| entry);
let expected = index_status.map(|index_status| {
TrackedStatus {
index_status,

View file

@ -204,10 +204,7 @@ fn assign_edit_prediction_provider(
}
}
let workspace = window
.root::<Workspace>()
.flatten()
.map(|workspace| workspace.downgrade());
let workspace = window.root::<Workspace>().flatten();
let zeta =
zeta::Zeta::register(workspace, worktree, client.clone(), user_store, cx);

View file

@ -6,19 +6,21 @@ mod onboarding_modal;
mod onboarding_telemetry;
mod rate_completion_modal;
use arrayvec::ArrayVec;
pub(crate) use completion_diff_element::*;
use db::kvp::{Dismissable, KEY_VALUE_STORE};
use edit_prediction::DataCollectionState;
pub use init::*;
use license_detection::LicenseDetectionWatcher;
use project::git_store::Repository;
pub use rate_completion_modal::*;
use anyhow::{Context as _, Result, anyhow};
use arrayvec::ArrayVec;
use client::{Client, EditPredictionUsage, UserStore};
use cloud_llm_client::{
AcceptEditPredictionBody, EXPIRED_LLM_TOKEN_HEADER_NAME, MINIMUM_REQUIRED_VERSION_HEADER_NAME,
PredictEditsBody, PredictEditsGitInfo, PredictEditsResponse, ZED_VERSION_HEADER_NAME,
PredictEditsBody, PredictEditsGitInfo, PredictEditsRecentFile, PredictEditsResponse,
ZED_VERSION_HEADER_NAME,
};
use collections::{HashMap, HashSet, VecDeque};
use futures::AsyncReadExt;
@ -32,7 +34,7 @@ use language::{
Anchor, Buffer, BufferSnapshot, EditPreview, OffsetRangeExt, ToOffset, ToPoint, text_diff,
};
use language_model::{LlmApiToken, RefreshLlmTokenListener};
use project::{Project, ProjectPath};
use project::{Project, ProjectEntryId, ProjectPath};
use release_channel::AppVersion;
use settings::WorktreeId;
use std::str::FromStr;
@ -70,6 +72,15 @@ const MAX_DIAGNOSTIC_GROUPS: usize = 10;
/// Maximum number of events to track.
const MAX_EVENT_COUNT: usize = 16;
/// Maximum number of recent files to track.
const MAX_RECENT_PROJECT_ENTRIES_COUNT: usize = 16;
/// Maximum file path length to include in recent files list.
const MAX_RECENT_FILE_PATH_LENGTH: usize = 512;
/// Maximum number of edit predictions to store for feedback.
const MAX_SHOWN_COMPLETION_COUNT: usize = 50;
actions!(
edit_prediction,
[
@ -212,7 +223,7 @@ impl std::fmt::Debug for EditPrediction {
}
pub struct Zeta {
workspace: Option<WeakEntity<Workspace>>,
workspace: WeakEntity<Workspace>,
client: Arc<Client>,
events: VecDeque<Event>,
registered_buffers: HashMap<gpui::EntityId, RegisteredBuffer>,
@ -225,6 +236,7 @@ pub struct Zeta {
update_required: bool,
user_store: Entity<UserStore>,
license_detection_watchers: HashMap<WorktreeId, Rc<LicenseDetectionWatcher>>,
recent_project_entries: VecDeque<(ProjectEntryId, Instant)>,
}
impl Zeta {
@ -233,7 +245,7 @@ impl Zeta {
}
pub fn register(
workspace: Option<WeakEntity<Workspace>>,
workspace: Option<Entity<Workspace>>,
worktree: Option<Entity<Worktree>>,
client: Arc<Client>,
user_store: Entity<UserStore>,
@ -266,7 +278,7 @@ impl Zeta {
}
fn new(
workspace: Option<WeakEntity<Workspace>>,
workspace: Option<Entity<Workspace>>,
client: Arc<Client>,
user_store: Entity<UserStore>,
cx: &mut Context<Self>,
@ -276,11 +288,27 @@ impl Zeta {
let data_collection_choice = Self::load_data_collection_choices();
let data_collection_choice = cx.new(|_| data_collection_choice);
if let Some(workspace) = &workspace {
cx.subscribe(
&workspace.read(cx).project().clone(),
|this, _workspace, event, _cx| match event {
project::Event::ActiveEntryChanged(Some(project_entry_id)) => {
this.push_recent_project_entry(*project_entry_id)
}
_ => {}
},
)
.detach();
}
Self {
workspace,
workspace: workspace.map_or_else(
|| WeakEntity::new_invalid(),
|workspace| workspace.downgrade(),
),
client,
events: VecDeque::new(),
shown_completions: VecDeque::new(),
events: VecDeque::with_capacity(MAX_EVENT_COUNT),
shown_completions: VecDeque::with_capacity(MAX_SHOWN_COMPLETION_COUNT),
rated_completions: HashSet::default(),
registered_buffers: HashMap::default(),
data_collection_choice,
@ -300,6 +328,7 @@ impl Zeta {
update_required: false,
license_detection_watchers: HashMap::default(),
user_store,
recent_project_entries: VecDeque::with_capacity(MAX_RECENT_PROJECT_ENTRIES_COUNT),
}
}
@ -327,11 +356,12 @@ impl Zeta {
}
}
self.events.push_back(event);
if self.events.len() >= MAX_EVENT_COUNT {
// These are halved instead of popping to improve prompt caching.
self.events.drain(..MAX_EVENT_COUNT / 2);
}
self.events.push_back(event);
}
pub fn register_buffer(&mut self, buffer: &Entity<Buffer>, cx: &mut Context<Self>) {
@ -374,7 +404,7 @@ impl Zeta {
project: Option<&Entity<Project>>,
buffer: &Entity<Buffer>,
cursor: language::Anchor,
can_collect_data: bool,
can_collect_data: CanCollectData,
cx: &mut Context<Self>,
perform_predict_edits: F,
) -> Task<Result<Option<EditPrediction>>>
@ -393,10 +423,8 @@ impl Zeta {
let llm_token = self.llm_token.clone();
let app_version = AppVersion::global(cx);
let git_info = if let (true, Some(project), Some(file)) =
(can_collect_data, project, snapshot.file())
{
git_info_for_file(project, &ProjectPath::from_file(file.as_ref(), cx), cx)
let git_info = if matches!(can_collect_data, CanCollectData(true)) {
self.gather_git_info(project.clone(), &buffer_snapshotted_at, &snapshot, cx)
} else {
None
};
@ -689,9 +717,15 @@ and then another
) -> Task<Result<Option<EditPrediction>>> {
use std::future::ready;
self.request_completion_impl(None, project, buffer, position, false, cx, |_params| {
ready(Ok((response, None)))
})
self.request_completion_impl(
None,
project,
buffer,
position,
CanCollectData(false),
cx,
|_params| ready(Ok((response, None))),
)
}
pub fn request_completion(
@ -699,15 +733,11 @@ and then another
project: Option<&Entity<Project>>,
buffer: &Entity<Buffer>,
position: language::Anchor,
can_collect_data: bool,
can_collect_data: CanCollectData,
cx: &mut Context<Self>,
) -> Task<Result<Option<EditPrediction>>> {
let workspace = self
.workspace
.as_ref()
.and_then(|workspace| workspace.upgrade());
self.request_completion_impl(
workspace,
self.workspace.upgrade(),
project,
buffer,
position,
@ -1021,11 +1051,11 @@ and then another
}
pub fn completion_shown(&mut self, completion: &EditPrediction, cx: &mut Context<Self>) {
self.shown_completions.push_front(completion.clone());
if self.shown_completions.len() > 50 {
if self.shown_completions.len() >= MAX_SHOWN_COMPLETION_COUNT {
let completion = self.shown_completions.pop_back().unwrap();
self.rated_completions.remove(&completion.id);
}
self.shown_completions.push_front(completion.clone());
cx.notify();
}
@ -1099,6 +1129,128 @@ and then another
None => DataCollectionChoice::NotAnswered,
}
}
fn gather_git_info(
&mut self,
project: Option<&Entity<Project>>,
buffer_snapshotted_at: &Instant,
snapshot: &BufferSnapshot,
cx: &Context<Self>,
) -> Option<PredictEditsGitInfo> {
let project = project?.read(cx);
let file = snapshot.file()?;
let project_path = ProjectPath::from_file(file.as_ref(), cx);
let entry = project.entry_for_path(&project_path, cx)?;
if !worktree_entry_eligible_for_collection(&entry) {
return None;
}
let git_store = project.git_store().read(cx);
let (repository, repo_path) =
git_store.repository_and_path_for_project_path(&project_path, cx)?;
let repo_path_str = repo_path.to_str()?;
let repository = repository.read(cx);
let head_sha = repository
.head_commit
.as_ref()
.map(|head_commit| head_commit.sha.to_string());
let remote_origin_url = repository.remote_origin_url.clone();
let remote_upstream_url = repository.remote_upstream_url.clone();
if head_sha.is_none() && remote_origin_url.is_none() && remote_upstream_url.is_none() {
return None;
}
let recent_files = self.recent_files(&buffer_snapshotted_at, repository, cx);
Some(PredictEditsGitInfo {
input_file_path: Some(repo_path_str.to_string()),
head_sha,
remote_origin_url,
remote_upstream_url,
recent_files: Some(recent_files),
})
}
fn push_recent_project_entry(&mut self, project_entry_id: ProjectEntryId) {
let now = Instant::now();
if let Some(existing_ix) = self
.recent_project_entries
.iter()
.rposition(|(id, _)| *id == project_entry_id)
{
self.recent_project_entries.remove(existing_ix);
}
if self.recent_project_entries.len() >= MAX_RECENT_PROJECT_ENTRIES_COUNT {
self.recent_project_entries.pop_front();
}
self.recent_project_entries
.push_back((project_entry_id, now));
}
fn recent_files(
&mut self,
now: &Instant,
repository: &Repository,
cx: &Context<Self>,
) -> Vec<PredictEditsRecentFile> {
let Ok(project) = self
.workspace
.read_with(cx, |workspace, _cx| workspace.project().clone())
else {
return Vec::new();
};
let mut results = Vec::new();
for ix in (0..self.recent_project_entries.len()).rev() {
let (entry_id, last_active_at) = &self.recent_project_entries[ix];
if let Some(worktree) = project.read(cx).worktree_for_entry(*entry_id, cx)
&& let worktree = worktree.read(cx)
&& let Some(entry) = worktree.entry_for_id(*entry_id)
&& worktree_entry_eligible_for_collection(entry)
{
let project_path = ProjectPath {
worktree_id: worktree.id(),
path: entry.path.clone(),
};
let Some(repo_path) = repository.project_path_to_repo_path(&project_path, cx)
else {
// entry not removed since queries involving other repositories might occur later
continue;
};
let Some(repo_path_str) = repo_path.to_str() else {
// paths may not be valid UTF-8
self.recent_project_entries.remove(ix);
continue;
};
if repo_path_str.len() > MAX_RECENT_FILE_PATH_LENGTH {
self.recent_project_entries.remove(ix);
continue;
}
let Ok(active_to_now_ms) =
now.duration_since(*last_active_at).as_millis().try_into()
else {
self.recent_project_entries.remove(ix);
continue;
};
results.push(PredictEditsRecentFile {
path: repo_path_str.to_string(),
active_to_now_ms,
});
} else {
self.recent_project_entries.remove(ix);
}
}
results
}
}
fn worktree_entry_eligible_for_collection(entry: &worktree::Entry) -> bool {
entry.is_file()
&& entry.is_created()
&& !entry.is_ignored
&& !entry.is_private
&& !entry.is_external
&& !entry.is_fifo
}
pub struct PerformPredictEditsParams {
@ -1123,35 +1275,6 @@ fn common_prefix<T1: Iterator<Item = char>, T2: Iterator<Item = char>>(a: T1, b:
.sum()
}
fn git_info_for_file(
project: &Entity<Project>,
project_path: &ProjectPath,
cx: &App,
) -> Option<PredictEditsGitInfo> {
let git_store = project.read(cx).git_store().read(cx);
if let Some((repository, _repo_path)) =
git_store.repository_and_path_for_project_path(project_path, cx)
{
let repository = repository.read(cx);
let head_sha = repository
.head_commit
.as_ref()
.map(|head_commit| head_commit.sha.to_string());
let remote_origin_url = repository.remote_origin_url.clone();
let remote_upstream_url = repository.remote_upstream_url.clone();
if head_sha.is_none() && remote_origin_url.is_none() && remote_upstream_url.is_none() {
return None;
}
Some(PredictEditsGitInfo {
head_sha,
remote_origin_url,
remote_upstream_url,
})
} else {
None
}
}
pub struct GatherContextOutput {
pub body: PredictEditsBody,
pub editable_range: Range<usize>,
@ -1163,14 +1286,16 @@ pub fn gather_context(
snapshot: &BufferSnapshot,
cursor_point: language::Point,
make_events_prompt: impl FnOnce() -> String + Send + 'static,
can_collect_data: bool,
can_collect_data: CanCollectData,
git_info: Option<PredictEditsGitInfo>,
cx: &App,
) -> Task<Result<GatherContextOutput>> {
let local_lsp_store =
project.and_then(|project| project.read(cx).lsp_store().read(cx).as_local());
let diagnostic_groups: Vec<(String, serde_json::Value)> =
if can_collect_data && let Some(local_lsp_store) = local_lsp_store {
if matches!(can_collect_data, CanCollectData(true))
&& let Some(local_lsp_store) = local_lsp_store
{
snapshot
.diagnostic_groups(None)
.into_iter()
@ -1211,7 +1336,7 @@ pub fn gather_context(
let body = PredictEditsBody {
input_events,
input_excerpt: input_excerpt.prompt,
can_collect_data,
can_collect_data: can_collect_data.0,
diagnostic_groups,
git_info,
outline: None,
@ -1376,6 +1501,9 @@ pub struct ProviderDataCollection {
license_detection_watcher: Option<Rc<LicenseDetectionWatcher>>,
}
#[derive(Debug, Clone, Copy)]
pub struct CanCollectData(pub bool);
impl ProviderDataCollection {
pub fn new(zeta: Entity<Zeta>, buffer: Option<Entity<Buffer>>, cx: &mut App) -> Self {
let choice_and_watcher = buffer.and_then(|buffer| {
@ -1409,8 +1537,8 @@ impl ProviderDataCollection {
}
}
pub fn can_collect_data(&self, cx: &App) -> bool {
self.is_data_collection_enabled(cx) && self.is_project_open_source()
pub fn can_collect_data(&self, cx: &App) -> CanCollectData {
CanCollectData(self.is_data_collection_enabled(cx) && self.is_project_open_source())
}
pub fn is_data_collection_enabled(&self, cx: &App) -> bool {
@ -2034,7 +2162,7 @@ mod tests {
let buffer = cx.new(|cx| Buffer::local(buffer_content, cx));
let cursor = buffer.read_with(cx, |buffer, _| buffer.anchor_before(Point::new(1, 0)));
let completion_task = zeta.update(cx, |zeta, cx| {
zeta.request_completion(None, &buffer, cursor, false, cx)
zeta.request_completion(None, &buffer, cursor, CanCollectData(false), cx)
});
let completion = completion_task.await.unwrap().unwrap();
@ -2099,7 +2227,7 @@ mod tests {
let snapshot = buffer.read_with(cx, |buffer, _| buffer.snapshot());
let cursor = buffer.read_with(cx, |buffer, _| buffer.anchor_before(Point::new(1, 0)));
let completion_task = zeta.update(cx, |zeta, cx| {
zeta.request_completion(None, &buffer, cursor, false, cx)
zeta.request_completion(None, &buffer, cursor, CanCollectData(false), cx)
});
let completion = completion_task.await.unwrap().unwrap();

View file

@ -18,7 +18,7 @@ use std::process::exit;
use std::str::FromStr;
use std::sync::Arc;
use std::time::Duration;
use zeta::{GatherContextOutput, PerformPredictEditsParams, Zeta, gather_context};
use zeta::{CanCollectData, GatherContextOutput, PerformPredictEditsParams, Zeta, gather_context};
use crate::headless::ZetaCliAppState;
@ -172,7 +172,6 @@ async fn get_context(
None => String::new(),
};
// Enable gathering extra data not currently needed for edit predictions
let can_collect_data = true;
let git_info = None;
let mut gather_context_output = cx
.update(|cx| {
@ -182,7 +181,7 @@ async fn get_context(
&snapshot,
clipped_cursor,
move || events,
can_collect_data,
CanCollectData(true),
git_info,
cx,
)