mod char_bag; mod fuzzy; use crate::{ editor::{History, Snapshot as BufferSnapshot}, sum_tree::{self, Edit, SumTree}, }; use anyhow::{anyhow, Result}; use fuzzy::PathEntry; pub use fuzzy::{match_paths, PathMatch}; use gpui::{scoped_pool, AppContext, Entity, ModelContext, ModelHandle, Task}; use ignore::dir::{Ignore, IgnoreBuilder}; use parking_lot::Mutex; use postage::{ prelude::{Sink, Stream}, watch, }; use smol::{channel::Sender, Timer}; use std::{ ffi::OsStr, fmt, fs, future::Future, io::{self, Read, Write}, ops::{AddAssign, Deref}, os::unix::fs::MetadataExt, path::{Path, PathBuf}, sync::Arc, time::Duration, }; #[derive(Clone, Debug)] enum ScanState { Idle, Scanning, Err(Arc), } pub struct Worktree { snapshot: Snapshot, scanner: Arc, scan_state: (watch::Sender, watch::Receiver), poll_scheduled: bool, } #[derive(Clone)] pub struct FileHandle { worktree: ModelHandle, inode: u64, } impl Worktree { pub fn new(path: impl Into>, ctx: &mut ModelContext) -> Self { let scan_state = smol::channel::unbounded(); let snapshot = Snapshot { id: ctx.model_id(), path: path.into(), root_inode: None, entries: Default::default(), }; let scanner = Arc::new(BackgroundScanner::new(snapshot.clone(), scan_state.0)); let tree = Self { snapshot, scanner, scan_state: watch::channel_with(ScanState::Scanning), poll_scheduled: false, }; let scanner = tree.scanner.clone(); std::thread::spawn(move || scanner.run()); ctx.spawn_stream(scan_state.1, Self::observe_scan_state, |_, _| {}) .detach(); tree } pub fn scan_complete(&self) -> impl Future { let mut scan_state_rx = self.scan_state.1.clone(); async move { let mut scan_state = Some(scan_state_rx.borrow().clone()); while let Some(ScanState::Scanning) = scan_state { scan_state = scan_state_rx.recv().await; } } } fn observe_scan_state(&mut self, scan_state: ScanState, ctx: &mut ModelContext) { let _ = self.scan_state.0.blocking_send(scan_state); self.poll_entries(ctx); } fn poll_entries(&mut self, ctx: &mut ModelContext) { self.snapshot = self.scanner.snapshot(); ctx.notify(); if self.is_scanning() && !self.poll_scheduled { ctx.spawn(Timer::after(Duration::from_millis(100)), |this, _, ctx| { this.poll_scheduled = false; this.poll_entries(ctx); }) .detach(); self.poll_scheduled = true; } } fn is_scanning(&self) -> bool { if let ScanState::Scanning = *self.scan_state.1.borrow() { true } else { false } } pub fn snapshot(&self) -> Snapshot { self.snapshot.clone() } pub fn contains_path(&self, path: &Path) -> bool { path.starts_with(&self.snapshot.path) } pub fn has_inode(&self, inode: u64) -> bool { self.snapshot.entries.get(&inode).is_some() } pub fn file_count(&self) -> usize { self.snapshot.entries.summary().file_count } pub fn abs_path_for_inode(&self, ino: u64) -> Result { let mut result = self.snapshot.path.to_path_buf(); result.push(self.path_for_inode(ino, false)?); Ok(result) } pub fn load_history( &self, ino: u64, ctx: &AppContext, ) -> impl Future> { let path = self.abs_path_for_inode(ino); ctx.background_executor().spawn(async move { let mut file = std::fs::File::open(&path?)?; let mut base_text = String::new(); file.read_to_string(&mut base_text)?; Ok(History::new(Arc::from(base_text))) }) } pub fn save<'a>( &self, ino: u64, content: BufferSnapshot, ctx: &AppContext, ) -> Task> { let path = self.abs_path_for_inode(ino); ctx.background_executor().spawn(async move { let buffer_size = content.text_summary().bytes.min(10 * 1024); let file = std::fs::File::create(&path?)?; let mut writer = std::io::BufWriter::with_capacity(buffer_size, file); for chunk in content.fragments() { writer.write(chunk.as_bytes())?; } writer.flush()?; Ok(()) }) } #[cfg(test)] pub fn files<'a>(&'a self) -> impl Iterator + 'a { self.snapshot .entries .cursor::<(), ()>() .filter_map(|entry| { if let Entry::File { inode, .. } = entry { Some(*inode) } else { None } }) } } impl Entity for Worktree { type Event = (); } impl Deref for Worktree { type Target = Snapshot; fn deref(&self) -> &Self::Target { &self.snapshot } } impl fmt::Debug for Worktree { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { self.snapshot.fmt(f) } } #[derive(Clone)] pub struct Snapshot { id: usize, path: Arc, root_inode: Option, entries: SumTree, } impl Snapshot { pub fn file_count(&self) -> usize { self.entries.summary().file_count } pub fn root_entry(&self) -> Option<&Entry> { self.root_inode.and_then(|inode| self.entries.get(&inode)) } pub fn root_name(&self) -> Option<&OsStr> { self.path.file_name() } fn entry_for_path(&self, path: impl AsRef) -> Option<&Entry> { self.inode_for_path(path) .and_then(|inode| self.entries.get(&inode)) } fn inode_for_path(&self, path: impl AsRef) -> Option { let path = path.as_ref(); self.root_inode.and_then(|mut inode| { 'components: for path_component in path { if let Some(Entry::Dir { children, .. }) = &self.entries.get(&inode) { for (child_inode, name) in children.as_ref() { if name.as_ref() == path_component { inode = *child_inode; continue 'components; } } } return None; } Some(inode) }) } pub fn path_for_inode(&self, mut inode: u64, include_root: bool) -> Result { let mut components = Vec::new(); let mut entry = self .entries .get(&inode) .ok_or_else(|| anyhow!("entry does not exist in worktree"))?; while let Some(parent) = entry.parent() { entry = self.entries.get(&parent).unwrap(); if let Entry::Dir { children, .. } = entry { let (_, child_name) = children .iter() .find(|(child_inode, _)| *child_inode == inode) .unwrap(); components.push(child_name.as_ref()); inode = parent; } else { unreachable!(); } } if include_root { components.push(self.root_name().unwrap()); } Ok(components.into_iter().rev().collect()) } fn insert_entry(&mut self, path: &Path, entry: Entry) { let mut edits = Vec::new(); edits.push(Edit::Insert(entry.clone())); if let Some(parent) = entry.parent() { if let Some(mut parent_entry) = self.entries.get(&parent).cloned() { if let Entry::Dir { children, .. } = &mut parent_entry { let name = Arc::from(path.file_name().unwrap()); *children = children .into_iter() .cloned() .chain(Some((entry.inode(), name))) .collect::>() .into(); edits.push(Edit::Insert(parent_entry)); } else { unreachable!(); } } } self.entries.edit(edits); } fn remove_entry(&mut self, path: &Path) { let mut parent_entry = match path.parent().and_then(|p| self.entry_for_path(p).cloned()) { Some(e) => e, None => return, }; let mut edits = Vec::new(); let parent_inode = parent_entry.inode(); let mut entry_inode = None; if let Entry::Dir { children, .. } = &mut parent_entry { let mut new_children = Vec::new(); for (child_inode, child_name) in children.as_ref() { if Some(child_name.as_ref()) == path.file_name() { entry_inode = Some(*child_inode); } else { new_children.push((*child_inode, child_name.clone())); } } if new_children.iter().any(|c| Some(c.0) == entry_inode) { entry_inode = None; } *children = new_children.into(); edits.push(Edit::Insert(parent_entry)); } if let Some(entry_inode) = entry_inode { let mut descendant_stack = Vec::new(); descendant_stack.push((entry_inode, parent_inode)); while let Some((inode, parent_inode)) = descendant_stack.pop() { if let Some(entry) = self.entries.get(&inode) { if entry.parent() == Some(parent_inode) { edits.push(Edit::Remove(inode)); if let Entry::Dir { children, .. } = entry { descendant_stack.extend(children.iter().map(|c| (c.0, entry.inode()))); } } } } } self.entries.edit(edits); } fn fmt_entry( &self, f: &mut fmt::Formatter<'_>, ino: u64, name: &OsStr, indent: usize, ) -> fmt::Result { match self.entries.get(&ino).unwrap() { Entry::Dir { children, .. } => { write!( f, "{}{}/ ({})\n", " ".repeat(indent), name.to_string_lossy(), ino )?; for (child_inode, child_name) in children.iter() { self.fmt_entry(f, *child_inode, child_name, indent + 2)?; } Ok(()) } Entry::File { .. } => write!( f, "{}{} ({})\n", " ".repeat(indent), name.to_string_lossy(), ino ), } } } impl fmt::Debug for Snapshot { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { if let Some(root_ino) = self.root_inode { self.fmt_entry(f, root_ino, self.path.file_name().unwrap(), 0) } else { write!(f, "Empty tree\n") } } } impl FileHandle { pub fn path(&self, ctx: &AppContext) -> PathBuf { self.worktree .read(ctx) .path_for_inode(self.inode, false) .unwrap() } pub fn load_history(&self, ctx: &AppContext) -> impl Future> { self.worktree.read(ctx).load_history(self.inode, ctx) } pub fn save<'a>(&self, content: BufferSnapshot, ctx: &AppContext) -> Task> { let worktree = self.worktree.read(ctx); worktree.save(self.inode, content, ctx) } pub fn entry_id(&self) -> (usize, u64) { (self.worktree.id(), self.inode) } } #[derive(Clone, Debug)] pub enum Entry { Dir { inode: u64, parent: Option, is_symlink: bool, is_ignored: bool, children: Arc<[(u64, Arc)]>, pending: bool, }, File { inode: u64, parent: Option, is_symlink: bool, is_ignored: bool, path: PathEntry, }, } impl Entry { fn inode(&self) -> u64 { match self { Entry::Dir { inode, .. } => *inode, Entry::File { inode, .. } => *inode, } } fn is_dir(&self) -> bool { matches!(self, Entry::Dir { .. }) } fn parent(&self) -> Option { match self { Entry::Dir { parent, .. } => *parent, Entry::File { parent, .. } => *parent, } } } impl sum_tree::Item for Entry { type Summary = EntrySummary; fn summary(&self) -> Self::Summary { EntrySummary { max_ino: self.inode(), file_count: if matches!(self, Self::File { .. }) { 1 } else { 0 }, } } } impl sum_tree::KeyedItem for Entry { type Key = u64; fn key(&self) -> Self::Key { self.inode() } } #[derive(Clone, Debug, Default)] pub struct EntrySummary { max_ino: u64, file_count: usize, } impl<'a> AddAssign<&'a EntrySummary> for EntrySummary { fn add_assign(&mut self, rhs: &'a EntrySummary) { self.max_ino = rhs.max_ino; self.file_count += rhs.file_count; } } impl<'a> sum_tree::Dimension<'a, EntrySummary> for u64 { fn add_summary(&mut self, summary: &'a EntrySummary) { *self = summary.max_ino; } } #[derive(Copy, Clone, Default, Debug, Eq, PartialEq, Ord, PartialOrd)] struct FileCount(usize); impl<'a> sum_tree::Dimension<'a, EntrySummary> for FileCount { fn add_summary(&mut self, summary: &'a EntrySummary) { self.0 += summary.file_count; } } struct BackgroundScanner { snapshot: Mutex, notify: Sender, thread_pool: scoped_pool::Pool, } impl BackgroundScanner { fn new(snapshot: Snapshot, notify: Sender) -> Self { Self { snapshot: Mutex::new(snapshot), notify, thread_pool: scoped_pool::Pool::new(16), } } fn path(&self) -> Arc { self.snapshot.lock().path.clone() } fn snapshot(&self) -> Snapshot { self.snapshot.lock().clone() } fn run(&self) { let path = self.snapshot.lock().path.clone(); // Create the event stream before we start scanning to ensure we receive events for changes // that occur in the middle of the scan. let event_stream = fsevent::EventStream::new(&[path.as_ref()], Duration::from_millis(100), |events| { if smol::block_on(self.notify.send(ScanState::Scanning)).is_err() { return false; } self.process_events(events); if smol::block_on(self.notify.send(ScanState::Idle)).is_err() { return false; } true }); if smol::block_on(self.notify.send(ScanState::Scanning)).is_err() { return; } if let Err(err) = self.scan_dirs() { if smol::block_on(self.notify.send(ScanState::Err(Arc::new(err)))).is_err() { return; } } if smol::block_on(self.notify.send(ScanState::Idle)).is_err() { return; } event_stream.run(); } fn scan_dirs(&self) -> io::Result<()> { let path = self.path(); let metadata = fs::metadata(&path)?; let inode = metadata.ino(); let is_symlink = fs::symlink_metadata(&path)?.file_type().is_symlink(); let name = Arc::from(path.file_name().unwrap_or(OsStr::new("/"))); let relative_path = PathBuf::from(&name); let mut ignore = IgnoreBuilder::new().build().add_parents(&path).unwrap(); if metadata.is_dir() { ignore = ignore.add_child(&path).unwrap(); } let is_ignored = ignore.matched(&path, metadata.is_dir()).is_ignore(); if metadata.file_type().is_dir() { let is_ignored = is_ignored || name.as_ref() == ".git"; let dir_entry = Entry::Dir { parent: None, inode, is_symlink, is_ignored, children: Arc::from([]), pending: true, }; self.insert_entries(Some(dir_entry.clone())); self.snapshot.lock().root_inode = Some(inode); let (tx, rx) = crossbeam_channel::unbounded(); tx.send(Ok(ScanJob { inode, path: path.clone(), relative_path, dir_entry, ignore: Some(ignore), scan_queue: tx.clone(), })) .unwrap(); drop(tx); let mut results = Vec::new(); results.resize_with(self.thread_pool.workers(), || Ok(())); self.thread_pool.scoped(|pool| { for result in &mut results { pool.execute(|| { let result = result; while let Ok(job) = rx.recv() { if let Err(err) = job.and_then(|job| self.scan_dir(job)) { *result = Err(err); break; } } }); } }); results.into_iter().collect::>()?; } else { self.insert_entries(Some(Entry::File { parent: None, path: PathEntry::new(inode, &relative_path, is_ignored), inode, is_symlink, is_ignored, })); self.snapshot.lock().root_inode = Some(inode); } Ok(()) } fn scan_dir(&self, job: ScanJob) -> io::Result<()> { let scan_queue = job.scan_queue; let mut dir_entry = job.dir_entry; let mut new_children = Vec::new(); let mut new_entries = Vec::new(); let mut new_jobs = Vec::new(); for child_entry in fs::read_dir(&job.path)? { let child_entry = child_entry?; let name: Arc = child_entry.file_name().into(); let relative_path = job.relative_path.join(name.as_ref()); let metadata = child_entry.metadata()?; let ino = metadata.ino(); let is_symlink = metadata.file_type().is_symlink(); let path = job.path.join(name.as_ref()); new_children.push((ino, name.clone())); if metadata.is_dir() { let mut is_ignored = true; let mut ignore = None; if let Some(parent_ignore) = job.ignore.as_ref() { let child_ignore = parent_ignore.add_child(&path).unwrap(); is_ignored = child_ignore.matched(&path, true).is_ignore() || name.as_ref() == ".git"; if !is_ignored { ignore = Some(child_ignore); } } let dir_entry = Entry::Dir { parent: Some(job.inode), inode: ino, is_symlink, is_ignored, children: Arc::from([]), pending: true, }; new_entries.push(dir_entry.clone()); new_jobs.push(ScanJob { inode: ino, path: Arc::from(path), relative_path, dir_entry, ignore, scan_queue: scan_queue.clone(), }); } else { let is_ignored = job .ignore .as_ref() .map_or(true, |i| i.matched(&path, false).is_ignore()); new_entries.push(Entry::File { parent: Some(job.inode), path: PathEntry::new(ino, &relative_path, is_ignored), inode: ino, is_symlink, is_ignored, }); }; } if let Entry::Dir { children, pending, .. } = &mut dir_entry { *children = Arc::from(new_children); *pending = false; } else { unreachable!() } new_entries.push(dir_entry); self.insert_entries(new_entries); for new_job in new_jobs { scan_queue.send(Ok(new_job)).unwrap(); } Ok(()) } fn process_events(&self, mut events: Vec) { let mut snapshot = self.snapshot(); let root_path = snapshot .path .canonicalize() .unwrap_or_else(|_| snapshot.path.to_path_buf()); events.sort_unstable_by(|a, b| a.path.cmp(&b.path)); let mut paths = events.into_iter().map(|e| e.path).peekable(); let (scan_queue_tx, scan_queue_rx) = crossbeam_channel::unbounded(); while let Some(path) = paths.next() { let relative_path = match path.strip_prefix(&root_path) { Ok(relative_path) => relative_path.to_path_buf(), Err(_) => { log::error!("unexpected event {:?} for root path {:?}", path, root_path); continue; } }; while paths.peek().map_or(false, |p| p.starts_with(&path)) { paths.next(); } snapshot.remove_entry(&relative_path); match self.fs_entry_for_path(&root_path, &path) { Ok(Some((fs_entry, ignore))) => { snapshot.insert_entry(&path, fs_entry.clone()); if fs_entry.is_dir() { scan_queue_tx .send(Ok(ScanJob { inode: fs_entry.inode(), path: Arc::from(path), relative_path: snapshot .root_name() .map_or(PathBuf::new(), PathBuf::from) .join(relative_path), dir_entry: fs_entry, ignore: Some(ignore), scan_queue: scan_queue_tx.clone(), })) .unwrap(); } } Ok(None) => {} Err(err) => { // TODO - create a special 'error' entry in the entries tree to mark this log::error!("error reading file on event {:?}", err); } } } *self.snapshot.lock() = snapshot; // Scan any directories that were created as part of this event batch. drop(scan_queue_tx); self.thread_pool.scoped(|pool| { for _ in 0..self.thread_pool.workers() { pool.execute(|| { while let Ok(job) = scan_queue_rx.recv() { if let Err(err) = job.and_then(|job| self.scan_dir(job)) { log::error!("Error scanning {:?}", err); } } }); } }); } fn fs_entry_for_path(&self, root_path: &Path, path: &Path) -> Result> { match fs::metadata(&path) { Ok(metadata) => { let mut ignore = IgnoreBuilder::new().build().add_parents(&path).unwrap(); if metadata.is_dir() { ignore = ignore.add_child(&path).unwrap(); } let is_ignored = ignore.matched(&path, metadata.is_dir()).is_ignore(); let inode = metadata.ino(); let is_symlink = fs::symlink_metadata(&path)?.file_type().is_symlink(); let parent = if path == root_path { None } else { Some(fs::metadata(path.parent().unwrap())?.ino()) }; if metadata.file_type().is_dir() { Ok(Some(( Entry::Dir { parent, inode, is_symlink, is_ignored, children: Arc::from([]), pending: true, }, ignore, ))) } else { Ok(Some(( Entry::File { parent, path: PathEntry::new( inode, root_path .parent() .map_or(path, |parent| path.strip_prefix(parent).unwrap()), is_ignored, ), inode, is_symlink, is_ignored, }, ignore, ))) } } Err(err) => { if err.kind() == io::ErrorKind::NotFound { Ok(None) } else { Err(anyhow::Error::new(err)) } } } } fn insert_entries(&self, entries: impl IntoIterator) { self.snapshot .lock() .entries .edit(entries.into_iter().map(Edit::Insert).collect::>()); } } impl Drop for BackgroundScanner { fn drop(&mut self) { self.thread_pool.shutdown(); } } struct ScanJob { inode: u64, path: Arc, relative_path: PathBuf, dir_entry: Entry, ignore: Option, scan_queue: crossbeam_channel::Sender>, } pub trait WorktreeHandle { fn file(&self, entry_id: u64, app: &AppContext) -> Result; } impl WorktreeHandle for ModelHandle { fn file(&self, inode: u64, app: &AppContext) -> Result { if self.read(app).has_inode(inode) { Ok(FileHandle { worktree: self.clone(), inode, }) } else { Err(anyhow!("entry does not exist in tree")) } } } trait UnwrapIgnoreTuple { fn unwrap(self) -> Ignore; } impl UnwrapIgnoreTuple for (Ignore, Option) { fn unwrap(self) -> Ignore { if let Some(error) = self.1 { log::error!("error loading gitignore data: {}", error); } self.0 } } #[cfg(test)] mod tests { use super::*; use crate::editor::Buffer; use crate::test::*; use anyhow::Result; use gpui::App; use rand::prelude::*; use serde_json::json; use std::env; use std::os::unix; use std::time::{SystemTime, UNIX_EPOCH}; #[test] fn test_populate_and_search() { App::test_async((), |mut app| async move { let dir = temp_tree(json!({ "root": { "apple": "", "banana": { "carrot": { "date": "", "endive": "", } }, "fennel": { "grape": "", } } })); let root_link_path = dir.path().join("root_link"); unix::fs::symlink(&dir.path().join("root"), &root_link_path).unwrap(); let tree = app.add_model(|ctx| Worktree::new(root_link_path, ctx)); app.read(|ctx| tree.read(ctx).scan_complete()).await; app.read(|ctx| { let tree = tree.read(ctx); assert_eq!(tree.file_count(), 4); let results = match_paths( Some(tree.snapshot()).iter(), "bna", false, false, false, 10, ctx.thread_pool().clone(), ) .iter() .map(|result| tree.path_for_inode(result.entry_id, true)) .collect::, _>>() .unwrap(); assert_eq!( results, vec![ PathBuf::from("root_link/banana/carrot/date"), PathBuf::from("root_link/banana/carrot/endive"), ] ); }) }); } #[test] fn test_save_file() { App::test_async((), |mut app| async move { let dir = temp_tree(json!({ "file1": "the old contents", })); let tree = app.add_model(|ctx| Worktree::new(dir.path(), ctx)); app.read(|ctx| tree.read(ctx).scan_complete()).await; app.read(|ctx| assert_eq!(tree.read(ctx).file_count(), 1)); let buffer = Buffer::new(1, "a line of text.\n".repeat(10 * 1024)); let file_inode = app.read(|ctx| { let tree = tree.read(ctx); let inode = tree.files().next().unwrap(); assert_eq!( tree.path_for_inode(inode, false) .unwrap() .file_name() .unwrap(), "file1" ); inode }); tree.update(&mut app, |tree, ctx| { smol::block_on(tree.save(file_inode, buffer.snapshot(), ctx.as_ref())).unwrap() }); let loaded_history = app .read(|ctx| tree.read(ctx).load_history(file_inode, ctx)) .await .unwrap(); assert_eq!(loaded_history.base_text.as_ref(), buffer.text()); }); } #[test] fn test_rescan() { App::test_async((), |mut app| async move { let dir = temp_tree(json!({ "a": { "file1": "", }, "b": { "c": { "file2": "", } } })); let tree = app.add_model(|ctx| Worktree::new(dir.path(), ctx)); app.read(|ctx| tree.read(ctx).scan_complete()).await; app.read(|ctx| assert_eq!(tree.read(ctx).file_count(), 2)); let file2 = app.read(|ctx| { let inode = tree.read(ctx).inode_for_path("b/c/file2").unwrap(); let file2 = tree.file(inode, ctx).unwrap(); assert_eq!(file2.path(ctx), Path::new("b/c/file2")); file2 }); std::fs::rename(dir.path().join("b/c"), dir.path().join("d")).unwrap(); tree.condition(&app, move |_, ctx| file2.path(ctx) == Path::new("d/file2")) .await; }); } #[test] fn test_random() { let iterations = env::var("ITERATIONS") .map(|i| i.parse().unwrap()) .unwrap_or(100); let operations = env::var("OPERATIONS") .map(|o| o.parse().unwrap()) .unwrap_or(40); let initial_entries = env::var("INITIAL_ENTRIES") .map(|o| o.parse().unwrap()) .unwrap_or(20); let seeds = if let Ok(seed) = env::var("SEED").map(|s| s.parse().unwrap()) { seed..seed + 1 } else { 0..iterations }; for seed in seeds { dbg!(seed); let mut rng = StdRng::seed_from_u64(seed); let root_dir = tempdir::TempDir::new(&format!("test-{}", seed)).unwrap(); for _ in 0..initial_entries { randomly_mutate_tree(root_dir.path(), 1.0, &mut rng).unwrap(); } log::info!("Generated initial tree"); let (notify_tx, _notify_rx) = smol::channel::unbounded(); let scanner = BackgroundScanner::new( Snapshot { id: 0, path: root_dir.path().into(), root_inode: None, entries: Default::default(), }, notify_tx, ); scanner.scan_dirs().unwrap(); let mut events = Vec::new(); let mut mutations_len = operations; while mutations_len > 1 { if !events.is_empty() && rng.gen_bool(0.4) { let len = rng.gen_range(0..=events.len()); let to_deliver = events.drain(0..len).collect::>(); log::info!("Delivering events: {:#?}", to_deliver); scanner.process_events(to_deliver); } else { events.extend(randomly_mutate_tree(root_dir.path(), 0.6, &mut rng).unwrap()); mutations_len -= 1; } } log::info!("Quiescing: {:#?}", events); scanner.process_events(events); let (notify_tx, _notify_rx) = smol::channel::unbounded(); let new_scanner = BackgroundScanner::new( Snapshot { id: 0, path: root_dir.path().into(), root_inode: None, entries: Default::default(), }, notify_tx, ); new_scanner.scan_dirs().unwrap(); assert_eq!(scanner.snapshot().to_vec(), new_scanner.snapshot().to_vec()); } } fn randomly_mutate_tree( root_path: &Path, insertion_probability: f64, rng: &mut impl Rng, ) -> Result> { let root_path = root_path.canonicalize().unwrap(); let (dirs, files) = read_dir_recursive(root_path.clone()); let mut events = Vec::new(); let mut record_event = |path: PathBuf| { events.push(fsevent::Event { event_id: SystemTime::now() .duration_since(UNIX_EPOCH) .unwrap() .as_secs(), flags: fsevent::StreamFlags::empty(), path, }); }; if (files.is_empty() && dirs.len() == 1) || rng.gen_bool(insertion_probability) { let path = dirs.choose(rng).unwrap(); let new_path = path.join(gen_name(rng)); if rng.gen() { log::info!("Creating dir {:?}", new_path.strip_prefix(root_path)?); fs::create_dir(&new_path)?; } else { log::info!("Creating file {:?}", new_path.strip_prefix(root_path)?); fs::write(&new_path, "")?; } record_event(new_path); } else { let old_path = { let file_path = files.choose(rng); let dir_path = dirs[1..].choose(rng); file_path.into_iter().chain(dir_path).choose(rng).unwrap() }; let is_rename = rng.gen(); if is_rename { let new_path_parent = dirs .iter() .filter(|d| !d.starts_with(old_path)) .choose(rng) .unwrap(); let overwrite_existing_dir = !old_path.starts_with(&new_path_parent) && rng.gen_bool(0.3); let new_path = if overwrite_existing_dir { fs::remove_dir_all(&new_path_parent).ok(); new_path_parent.to_path_buf() } else { new_path_parent.join(gen_name(rng)) }; log::info!( "Renaming {:?} to {}{:?}", old_path.strip_prefix(&root_path)?, if overwrite_existing_dir { "overwrite " } else { "" }, new_path.strip_prefix(&root_path)? ); fs::rename(&old_path, &new_path)?; record_event(old_path.clone()); record_event(new_path); } else if old_path.is_dir() { let (dirs, files) = read_dir_recursive(old_path.clone()); log::info!("Deleting dir {:?}", old_path.strip_prefix(&root_path)?); fs::remove_dir_all(&old_path).unwrap(); for file in files { record_event(file); } for dir in dirs { record_event(dir); } } else { log::info!("Deleting file {:?}", old_path.strip_prefix(&root_path)?); fs::remove_file(old_path).unwrap(); record_event(old_path.clone()); } } Ok(events) } fn read_dir_recursive(path: PathBuf) -> (Vec, Vec) { let child_entries = fs::read_dir(&path).unwrap(); let mut dirs = vec![path]; let mut files = Vec::new(); for child_entry in child_entries { let child_path = child_entry.unwrap().path(); if child_path.is_dir() { let (child_dirs, child_files) = read_dir_recursive(child_path); dirs.extend(child_dirs); files.extend(child_files); } else { files.push(child_path); } } (dirs, files) } fn gen_name(rng: &mut impl Rng) -> String { (0..6) .map(|_| rng.sample(rand::distributions::Alphanumeric)) .map(char::from) .collect() } impl Snapshot { fn to_vec(&self) -> Vec<(PathBuf, u64)> { use std::iter::FromIterator; let mut paths = Vec::new(); let mut stack = Vec::new(); stack.extend(self.root_inode); while let Some(inode) = stack.pop() { let computed_path = self.path_for_inode(inode, true).unwrap(); match self.entries.get(&inode).unwrap() { Entry::Dir { children, .. } => { stack.extend(children.iter().map(|c| c.0)); } Entry::File { path, .. } => { assert_eq!( String::from_iter(path.path.iter()), computed_path.to_str().unwrap() ); } } paths.push((computed_path, inode)); } paths.sort_by(|a, b| a.0.cmp(&b.0)); paths } } }