Converted to sqlez, so much nicer

This commit is contained in:
Mikayla Maki 2022-11-02 13:26:23 -07:00
parent c8face33fa
commit 406663c75e
12 changed files with 278 additions and 197 deletions

1
Cargo.lock generated
View file

@ -1559,6 +1559,7 @@ dependencies = [
"parking_lot 0.11.2", "parking_lot 0.11.2",
"sqlez", "sqlez",
"tempdir", "tempdir",
"util",
] ]
[[package]] [[package]]

View file

@ -11,11 +11,12 @@ doctest = false
test-support = [] test-support = []
[dependencies] [dependencies]
indoc = "1.0.4"
collections = { path = "../collections" } collections = { path = "../collections" }
gpui = { path = "../gpui" } gpui = { path = "../gpui" }
sqlez = { path = "../sqlez" } sqlez = { path = "../sqlez" }
util = { path = "../util" }
anyhow = "1.0.57" anyhow = "1.0.57"
indoc = "1.0.4"
async-trait = "0.1" async-trait = "0.1"
lazy_static = "1.4.0" lazy_static = "1.4.0"
log = { version = "0.4.16", features = ["kv_unstable_serde"] } log = { version = "0.4.16", features = ["kv_unstable_serde"] }

View file

@ -7,10 +7,8 @@ const TEST_FILE: &'static str = "test-db.db";
fn main() -> anyhow::Result<()> { fn main() -> anyhow::Result<()> {
env_logger::init(); env_logger::init();
let db = db::Db::open_in_memory(); let db = db::Db::open_in_memory("db");
if db.real().is_none() {
return Err(anyhow::anyhow!("Migrations failed"));
}
let file = Path::new(TEST_FILE); let file = Path::new(TEST_FILE);
let f = File::create(file)?; let f = File::create(file)?;
@ -21,21 +19,21 @@ fn main() -> anyhow::Result<()> {
let workspace_3 = db.workspace_for_roots(&["/tmp3", "/tmp2"]); let workspace_3 = db.workspace_for_roots(&["/tmp3", "/tmp2"]);
db.save_dock_pane( db.save_dock_pane(
workspace_1.workspace_id, &workspace_1.workspace_id,
&SerializedDockPane { &SerializedDockPane {
anchor_position: DockAnchor::Expanded, anchor_position: DockAnchor::Expanded,
visible: true, visible: true,
}, },
); );
db.save_dock_pane( db.save_dock_pane(
workspace_2.workspace_id, &workspace_2.workspace_id,
&SerializedDockPane { &SerializedDockPane {
anchor_position: DockAnchor::Bottom, anchor_position: DockAnchor::Bottom,
visible: true, visible: true,
}, },
); );
db.save_dock_pane( db.save_dock_pane(
workspace_3.workspace_id, &workspace_3.workspace_id,
&SerializedDockPane { &SerializedDockPane {
anchor_position: DockAnchor::Right, anchor_position: DockAnchor::Right,
visible: false, visible: false,

View file

@ -4,10 +4,8 @@ const TEST_FILE: &'static str = "test-db.db";
fn main() -> anyhow::Result<()> { fn main() -> anyhow::Result<()> {
env_logger::init(); env_logger::init();
let db = db::Db::open_in_memory(); let db = db::Db::open_in_memory("db");
if db.real().is_none() {
return Err(anyhow::anyhow!("Migrations failed"));
}
let file = Path::new(TEST_FILE); let file = Path::new(TEST_FILE);
let f = File::create(file)?; let f = File::create(file)?;

View file

@ -18,7 +18,7 @@ use sqlez::thread_safe_connection::ThreadSafeConnection;
pub use workspace::*; pub use workspace::*;
#[derive(Clone)] #[derive(Clone)]
struct Db(ThreadSafeConnection); pub struct Db(ThreadSafeConnection);
impl Deref for Db { impl Deref for Db {
type Target = sqlez::connection::Connection; type Target = sqlez::connection::Connection;
@ -54,15 +54,15 @@ impl Db {
} }
/// Open a in memory database for testing and as a fallback. /// Open a in memory database for testing and as a fallback.
pub fn open_in_memory() -> Self { pub fn open_in_memory(db_name: &str) -> Self {
Db( Db(ThreadSafeConnection::new(db_name, false)
ThreadSafeConnection::new("Zed DB", false).with_initialize_query(indoc! {" .with_initialize_query(indoc! {"
PRAGMA journal_mode=WAL; PRAGMA journal_mode=WAL;
PRAGMA synchronous=NORMAL; PRAGMA synchronous=NORMAL;
PRAGMA foreign_keys=TRUE; PRAGMA foreign_keys=TRUE;
PRAGMA case_sensitive_like=TRUE; PRAGMA case_sensitive_like=TRUE;
"}), "})
) .with_migrations(&[KVP_MIGRATION, WORKSPACES_MIGRATION, PANE_MIGRATIONS]))
} }
pub fn write_to<P: AsRef<Path>>(&self, dest: P) -> Result<()> { pub fn write_to<P: AsRef<Path>>(&self, dest: P) -> Result<()> {

View file

@ -23,7 +23,7 @@ impl Db {
pub fn write_kvp(&self, key: &str, value: &str) -> Result<()> { pub fn write_kvp(&self, key: &str, value: &str) -> Result<()> {
self.0 self.0
.prepare("INSERT OR REPLACE INTO kv_store(key, value) VALUES (?, ?)")? .prepare("INSERT OR REPLACE INTO kv_store(key, value) VALUES ((?), (?))")?
.with_bindings((key, value))? .with_bindings((key, value))?
.exec() .exec()
} }
@ -44,21 +44,21 @@ mod tests {
#[test] #[test]
fn test_kvp() -> Result<()> { fn test_kvp() -> Result<()> {
let db = Db::open_in_memory(); let db = Db::open_in_memory("test_kvp");
assert_eq!(db.read_kvp("key-1")?, None); assert_eq!(db.read_kvp("key-1").unwrap(), None);
db.write_kvp("key-1", "one")?; db.write_kvp("key-1", "one").unwrap();
assert_eq!(db.read_kvp("key-1")?, Some("one".to_string())); assert_eq!(db.read_kvp("key-1").unwrap(), Some("one".to_string()));
db.write_kvp("key-1", "one-2")?; db.write_kvp("key-1", "one-2").unwrap();
assert_eq!(db.read_kvp("key-1")?, Some("one-2".to_string())); assert_eq!(db.read_kvp("key-1").unwrap(), Some("one-2".to_string()));
db.write_kvp("key-2", "two")?; db.write_kvp("key-2", "two").unwrap();
assert_eq!(db.read_kvp("key-2")?, Some("two".to_string())); assert_eq!(db.read_kvp("key-2").unwrap(), Some("two".to_string()));
db.delete_kvp("key-1")?; db.delete_kvp("key-1").unwrap();
assert_eq!(db.read_kvp("key-1")?, None); assert_eq!(db.read_kvp("key-1").unwrap(), None);
Ok(()) Ok(())
} }

View file

@ -1,16 +1,21 @@
use std::str::FromStr; use std::str::FromStr;
use gpui::Axis; use gpui::Axis;
use indoc::indoc; use indoc::indoc;
use sqlez::{migrations::Migration, bindable::{Bind, Column}, connection::Connection, statement::Statement}; use sqlez::{
bindable::{Bind, Column},
migrations::Migration,
statement::Statement,
};
use util::{iife, ResultExt};
use crate::{items::ItemId, workspace::WorkspaceId}; use crate::{items::ItemId, workspace::WorkspaceId};
use super::Db; use super::Db;
pub(crate) const PANE_MIGRATIONS: Migration = Migration::new("pane", &[indoc! {" pub(crate) const PANE_MIGRATIONS: Migration = Migration::new(
"pane",
&[indoc! {"
CREATE TABLE dock_panes( CREATE TABLE dock_panes(
dock_pane_id INTEGER PRIMARY KEY, dock_pane_id INTEGER PRIMARY KEY,
workspace_id INTEGER NOT NULL, workspace_id INTEGER NOT NULL,
@ -19,7 +24,7 @@ CREATE TABLE dock_panes(
FOREIGN KEY(workspace_id) REFERENCES workspaces(workspace_id) ON DELETE CASCADE FOREIGN KEY(workspace_id) REFERENCES workspaces(workspace_id) ON DELETE CASCADE
) STRICT; ) STRICT;
CREATE TABLE pane_groups( CREATE TABLE pane_groups( -- Inner nodes
group_id INTEGER PRIMARY KEY, group_id INTEGER PRIMARY KEY,
workspace_id INTEGER NOT NULL, workspace_id INTEGER NOT NULL,
parent_group INTEGER, -- NULL indicates that this is a root node parent_group INTEGER, -- NULL indicates that this is a root node
@ -28,7 +33,8 @@ CREATE TABLE pane_groups(
FOREIGN KEY(parent_group) REFERENCES pane_groups(group_id) ON DELETE CASCADE FOREIGN KEY(parent_group) REFERENCES pane_groups(group_id) ON DELETE CASCADE
) STRICT; ) STRICT;
CREATE TABLE grouped_panes(
CREATE TABLE grouped_panes( -- Leaf nodes
pane_id INTEGER PRIMARY KEY, pane_id INTEGER PRIMARY KEY,
workspace_id INTEGER NOT NULL, workspace_id INTEGER NOT NULL,
group_id INTEGER NOT NULL, group_id INTEGER NOT NULL,
@ -65,7 +71,8 @@ CREATE TABLE dock_items(
FOREIGN KEY(dock_pane_id) REFERENCES dock_panes(dock_pane_id) ON DELETE CASCADE, FOREIGN KEY(dock_pane_id) REFERENCES dock_panes(dock_pane_id) ON DELETE CASCADE,
FOREIGN KEY(item_id) REFERENCES items(item_id)ON DELETE CASCADE FOREIGN KEY(item_id) REFERENCES items(item_id)ON DELETE CASCADE
) STRICT; ) STRICT;
"}]); "}],
);
// We have an many-branched, unbalanced tree with three types: // We have an many-branched, unbalanced tree with three types:
// Pane Groups // Pane Groups
@ -137,10 +144,9 @@ pub struct SerializedPane {
children: Vec<ItemId>, children: Vec<ItemId>,
} }
//********* CURRENTLY IN USE TYPES: ********* //********* CURRENTLY IN USE TYPES: *********
#[derive(Default, Debug, PartialEq, Eq)] #[derive(Default, Debug, PartialEq, Eq, Clone, Copy)]
pub enum DockAnchor { pub enum DockAnchor {
#[default] #[default]
Bottom, Bottom,
@ -162,15 +168,28 @@ impl FromStr for DockAnchor {
type Err = anyhow::Error; type Err = anyhow::Error;
fn from_str(s: &str) -> anyhow::Result<Self> { fn from_str(s: &str) -> anyhow::Result<Self> {
match s { match s {
"Bottom" => Ok(DockAnchor::Bottom), "Bottom" => Ok(DockAnchor::Bottom),
"Right" => Ok(DockAnchor::Right), "Right" => Ok(DockAnchor::Right),
"Expanded" => Ok(DockAnchor::Expanded), "Expanded" => Ok(DockAnchor::Expanded),
_ => anyhow::bail!("Not a valid dock anchor") _ => anyhow::bail!("Not a valid dock anchor"),
} }
} }
} }
impl Bind for DockAnchor {
fn bind(&self, statement: &Statement, start_index: i32) -> anyhow::Result<i32> {
statement.bind(self.to_string(), start_index)
}
}
impl Column for DockAnchor {
fn column(statement: &mut Statement, start_index: i32) -> anyhow::Result<(Self, i32)> {
<String as Column>::column(statement, start_index)
.and_then(|(str, next_index)| Ok((DockAnchor::from_str(&str)?, next_index)))
}
}
#[derive(Default, Debug, PartialEq, Eq)] #[derive(Default, Debug, PartialEq, Eq)]
pub struct SerializedDockPane { pub struct SerializedDockPane {
pub anchor_position: DockAnchor, pub anchor_position: DockAnchor,
@ -178,11 +197,30 @@ pub struct SerializedDockPane {
} }
impl SerializedDockPane { impl SerializedDockPane {
pub fn to_row(&self, workspace: WorkspaceId) -> DockRow { fn to_row(&self, workspace: &WorkspaceId) -> DockRow {
DockRow { workspace_id: workspace, anchor_position: self.anchor_position, visible: self.visible } DockRow {
workspace_id: *workspace,
anchor_position: self.anchor_position,
visible: self.visible,
}
} }
} }
impl Column for SerializedDockPane {
fn column(statement: &mut Statement, start_index: i32) -> anyhow::Result<(Self, i32)> {
<(DockAnchor, bool) as Column>::column(statement, start_index).map(
|((anchor_position, visible), next_index)| {
(
SerializedDockPane {
anchor_position,
visible,
},
next_index,
)
},
)
}
}
#[derive(Default, Debug, PartialEq, Eq)] #[derive(Default, Debug, PartialEq, Eq)]
pub(crate) struct DockRow { pub(crate) struct DockRow {
@ -191,24 +229,16 @@ pub(crate) struct DockRow {
visible: bool, visible: bool,
} }
impl DockRow {
pub fn to_pane(&self) -> SerializedDockPane {
SerializedDockPane { anchor_position: self.anchor_position, visible: self.visible }
}
}
impl Bind for DockRow { impl Bind for DockRow {
fn bind(&self, statement: &Statement, start_index: i32) -> anyhow::Result<i32> { fn bind(&self, statement: &Statement, start_index: i32) -> anyhow::Result<i32> {
statement.bind((self.workspace_id, self.anchor_position.to_string(), self.visible), start_index) statement.bind(
} (
} self.workspace_id,
self.anchor_position.to_string(),
impl Column for DockRow { self.visible,
fn column(statement: &mut Statement, start_index: i32) -> anyhow::Result<(Self, i32)> { ),
<(WorkspaceId, &str, bool) as Column>::column(statement, start_index) start_index,
.map(|((workspace_id, anchor_position, visible), next_index)| { )
})
} }
} }
@ -267,75 +297,37 @@ impl Db {
} }
pub fn get_dock_pane(&self, workspace: WorkspaceId) -> Option<SerializedDockPane> { pub fn get_dock_pane(&self, workspace: WorkspaceId) -> Option<SerializedDockPane> {
fn logic(conn: &Connection, workspace: WorkspaceId) -> anyhow::Result<Option<SerializedDockPane>> { iife!({
self.prepare("SELECT anchor_position, visible FROM dock_panes WHERE workspace_id = ?")?
let mut stmt = conn.prepare("SELECT workspace_id, anchor_position, visible FROM dock_panes WHERE workspace_id = ?")? .with_bindings(workspace)?
.maybe_row() .maybe_row::<SerializedDockPane>()
.map(|row| DockRow::col); })
.log_err()
.flatten()
let dock_panes = stmt.query_row([workspace.raw_id()], |row_ref| from_row::<DockRow>).optional();
let mut dock_panes_iter = stmt.query_and_then([workspace.raw_id()], from_row::<DockRow>)?;
let dock_pane = dock_panes_iter
.next()
.and_then(|dock_row|
dock_row
.ok()
.map(|dock_row| dock_row.to_pane()));
Ok(dock_pane)
}
self.real()
.map(|db| {
let lock = db.connection.lock();
match logic(&lock, workspace) {
Ok(dock_pane) => dock_pane,
Err(err) => {
log::error!("Failed to get the dock pane: {}", err);
None
},
}
})
.unwrap_or(None)
} }
pub fn save_dock_pane(&self, workspace: WorkspaceId, dock_pane: SerializedDockPane) { pub fn save_dock_pane(&self, workspace: &WorkspaceId, dock_pane: &SerializedDockPane) {
to_params_named(dock_pane.to_row(workspace)) iife!({
.map_err(|err| { self.prepare(
log::error!("Failed to parse params for the dock row: {}", err); "INSERT INTO dock_panes (workspace_id, anchor_position, visible) VALUES (?, ?, ?);",
err )?
}) .with_bindings(dock_pane.to_row(workspace))?
.ok() .insert()
.zip(self.real()) })
.map(|(params, db)| { .log_err();
// TODO: overwrite old dock panes if need be
let query = "INSERT INTO dock_panes (workspace_id, anchor_position, visible) VALUES (:workspace_id, :anchor_position, :visible);";
db.connection
.lock()
.execute(query, params.to_slice().as_slice())
.map(|_| ()) // Eat the return value
.unwrap_or_else(|err| {
log::error!("Failed to insert new dock pane into DB: {}", err);
})
});
} }
} }
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use crate::Db; use crate::{pane::SerializedPane, Db};
use super::{DockAnchor, SerializedDockPane}; use super::{DockAnchor, SerializedDockPane};
#[test] #[test]
fn test_basic_dock_pane() { fn test_basic_dock_pane() {
let db = Db::open_in_memory(); let db = Db::open_in_memory("basic_dock_pane");
let workspace = db.workspace_for_roots(&["/tmp"]); let workspace = db.workspace_for_roots(&["/tmp"]);
@ -344,7 +336,28 @@ mod tests {
visible: true, visible: true,
}; };
db.save_dock_pane(workspace.workspace_id, dock_pane); db.save_dock_pane(&workspace.workspace_id, &dock_pane);
let new_workspace = db.workspace_for_roots(&["/tmp"]);
assert_eq!(new_workspace.dock_pane.unwrap(), dock_pane);
}
#[test]
fn test_dock_simple_split() {
let db = Db::open_in_memory("simple_split");
let workspace = db.workspace_for_roots(&["/tmp"]);
let center_pane = SerializedPane {
pane_id: crate::pane::PaneId {
workspace_id: workspace.workspace_id,
pane_id: 1,
},
children: vec![],
};
db.save_dock_pane(&workspace.workspace_id, &dock_pane);
let new_workspace = db.workspace_for_roots(&["/tmp"]); let new_workspace = db.workspace_for_roots(&["/tmp"]);

View file

@ -1,4 +1,4 @@
use anyhow::{Result, anyhow}; use anyhow::Result;
use std::{ use std::{
ffi::OsStr, ffi::OsStr,
@ -10,7 +10,9 @@ use std::{
use indoc::indoc; use indoc::indoc;
use sqlez::{ use sqlez::{
connection::Connection, migrations::Migration, bindable::{Column, Bind}, bindable::{Bind, Column},
connection::Connection,
migrations::Migration,
}; };
use crate::pane::SerializedDockPane; use crate::pane::SerializedDockPane;
@ -47,13 +49,17 @@ impl WorkspaceId {
impl Bind for WorkspaceId { impl Bind for WorkspaceId {
fn bind(&self, statement: &sqlez::statement::Statement, start_index: i32) -> Result<i32> { fn bind(&self, statement: &sqlez::statement::Statement, start_index: i32) -> Result<i32> {
todo!(); statement.bind(self.raw_id(), start_index)
} }
} }
impl Column for WorkspaceId { impl Column for WorkspaceId {
fn column(statement: &mut sqlez::statement::Statement, start_index: i32) -> Result<(Self, i32)> { fn column(
todo!(); statement: &mut sqlez::statement::Statement,
start_index: i32,
) -> Result<(Self, i32)> {
<i64 as Column>::column(statement, start_index)
.map(|(id, next_index)| (WorkspaceId(id), next_index))
} }
} }
@ -154,10 +160,8 @@ impl Db {
fn last_workspace_id(&self) -> Option<WorkspaceId> { fn last_workspace_id(&self) -> Option<WorkspaceId> {
let res = self let res = self
.prepare( .prepare("SELECT workspace_id FROM workspaces ORDER BY timestamp DESC LIMIT 1")
"SELECT workspace_id FROM workspaces ORDER BY last_opened_timestamp DESC LIMIT 1", .and_then(|mut stmt| stmt.maybe_row())
)
.and_then(|stmt| stmt.maybe_row())
.map(|row| row.map(|id| WorkspaceId(id))); .map(|row| row.map(|id| WorkspaceId(id)));
match res { match res {
@ -172,28 +176,30 @@ impl Db {
/// Returns the previous workspace ids sorted by last modified along with their opened worktree roots /// Returns the previous workspace ids sorted by last modified along with their opened worktree roots
pub fn recent_workspaces(&self, limit: usize) -> Vec<(WorkspaceId, Vec<Arc<Path>>)> { pub fn recent_workspaces(&self, limit: usize) -> Vec<(WorkspaceId, Vec<Arc<Path>>)> {
self.with_savepoint("recent_workspaces", |conn| { self.with_savepoint("recent_workspaces", |conn| {
let ids = conn.prepare("SELECT workspace_id FROM workspaces ORDER BY last_opened_timestamp DESC LIMIT ?")? let rows = conn
.prepare("SELECT workspace_id FROM workspaces ORDER BY timestamp DESC LIMIT ?")?
.with_bindings(limit)? .with_bindings(limit)?
.rows::<i64>()? .rows::<i64>()?;
.iter()
.map(|row| WorkspaceId(*row)); let ids = rows.iter().map(|row| WorkspaceId(*row));
let result = Vec::new(); let mut result = Vec::new();
let stmt = conn.prepare("SELECT worktree_root FROM worktree_roots WHERE workspace_id = ?")?; let mut stmt =
conn.prepare("SELECT worktree_root FROM worktree_roots WHERE workspace_id = ?")?;
for workspace_id in ids { for workspace_id in ids {
let roots = stmt.with_bindings(workspace_id.0)? let roots = stmt
.with_bindings(workspace_id.0)?
.rows::<Vec<u8>>()? .rows::<Vec<u8>>()?
.iter() .iter()
.map(|row| { .map(|row| PathBuf::from(OsStr::from_bytes(&row)).into())
PathBuf::from(OsStr::from_bytes(&row)).into()
})
.collect(); .collect();
result.push((workspace_id, roots)) result.push((workspace_id, roots))
} }
Ok(result) Ok(result)
}).unwrap_or_else(|err| { })
.unwrap_or_else(|err| {
log::error!("Failed to get recent workspaces, err: {}", err); log::error!("Failed to get recent workspaces, err: {}", err);
Vec::new() Vec::new()
}) })
@ -213,11 +219,10 @@ where
if let Some(preexisting_id) = preexisting_id { if let Some(preexisting_id) = preexisting_id {
if preexisting_id != *workspace_id { if preexisting_id != *workspace_id {
// Should also delete fields in other tables with cascading updates // Should also delete fields in other tables with cascading updates
connection.prepare( connection
"DELETE FROM workspaces WHERE workspace_id = ?", .prepare("DELETE FROM workspaces WHERE workspace_id = ?")?
)? .with_bindings(preexisting_id.0)?
.with_bindings(preexisting_id.0)? .exec()?;
.exec()?;
} }
} }
@ -231,12 +236,14 @@ where
// If you need to debug this, here's the string parsing: // If you need to debug this, here's the string parsing:
// let path = root.as_ref().to_string_lossy().to_string(); // let path = root.as_ref().to_string_lossy().to_string();
connection.prepare("INSERT INTO worktree_roots(workspace_id, worktree_root) VALUES (?, ?)")? connection
.prepare("INSERT INTO worktree_roots(workspace_id, worktree_root) VALUES (?, ?)")?
.with_bindings((workspace_id.0, path))? .with_bindings((workspace_id.0, path))?
.exec()?; .exec()?;
} }
connection.prepare("UPDATE workspaces SET last_opened_timestamp = CURRENT_TIMESTAMP WHERE workspace_id = ?")? connection
.prepare("UPDATE workspaces SET timestamp = CURRENT_TIMESTAMP WHERE workspace_id = ?")?
.with_bindings(workspace_id.0)? .with_bindings(workspace_id.0)?
.exec()?; .exec()?;
@ -264,7 +271,7 @@ where
} }
} }
array_binding_stmt.push(')'); array_binding_stmt.push(')');
// Any workspace can have multiple independent paths, and these paths // Any workspace can have multiple independent paths, and these paths
// can overlap in the database. Take this test data for example: // can overlap in the database. Take this test data for example:
// //
@ -336,10 +343,14 @@ where
// Make sure we bound the parameters correctly // Make sure we bound the parameters correctly
debug_assert!(worktree_roots.len() as i32 + 1 == stmt.parameter_count()); debug_assert!(worktree_roots.len() as i32 + 1 == stmt.parameter_count());
let root_bytes: Vec<&[u8]> = worktree_roots.iter() let root_bytes: Vec<&[u8]> = worktree_roots
.map(|root| root.as_ref().as_os_str().as_bytes()).collect(); .iter()
.map(|root| root.as_ref().as_os_str().as_bytes())
stmt.with_bindings((root_bytes, root_bytes.len()))? .collect();
let len = root_bytes.len();
stmt.with_bindings((root_bytes, len))?
.maybe_row() .maybe_row()
.map(|row| row.map(|id| WorkspaceId(id))) .map(|row| row.map(|id| WorkspaceId(id)))
} }
@ -360,7 +371,8 @@ mod tests {
#[test] #[test]
fn test_new_worktrees_for_roots() { fn test_new_worktrees_for_roots() {
let db = Db::open_in_memory(); env_logger::init();
let db = Db::open_in_memory("test_new_worktrees_for_roots");
// Test creation in 0 case // Test creation in 0 case
let workspace_1 = db.workspace_for_roots::<String>(&[]); let workspace_1 = db.workspace_for_roots::<String>(&[]);
@ -371,7 +383,7 @@ mod tests {
assert_eq!(workspace_1.workspace_id, WorkspaceId(1)); assert_eq!(workspace_1.workspace_id, WorkspaceId(1));
// Ensure the timestamps are different // Ensure the timestamps are different
sleep(Duration::from_millis(20)); sleep(Duration::from_secs(1));
db.make_new_workspace::<String>(&[]); db.make_new_workspace::<String>(&[]);
// Test pulling another value from recent workspaces // Test pulling another value from recent workspaces
@ -379,7 +391,7 @@ mod tests {
assert_eq!(workspace_2.workspace_id, WorkspaceId(2)); assert_eq!(workspace_2.workspace_id, WorkspaceId(2));
// Ensure the timestamps are different // Ensure the timestamps are different
sleep(Duration::from_millis(20)); sleep(Duration::from_secs(1));
// Test creating a new workspace that doesn't exist already // Test creating a new workspace that doesn't exist already
let workspace_3 = db.workspace_for_roots(&["/tmp", "/tmp2"]); let workspace_3 = db.workspace_for_roots(&["/tmp", "/tmp2"]);
@ -396,7 +408,7 @@ mod tests {
#[test] #[test]
fn test_empty_worktrees() { fn test_empty_worktrees() {
let db = Db::open_in_memory(); let db = Db::open_in_memory("test_empty_worktrees");
assert_eq!(None, db.workspace_id::<String>(&[])); assert_eq!(None, db.workspace_id::<String>(&[]));
@ -404,7 +416,6 @@ mod tests {
db.make_new_workspace::<String>(&[]); //ID 2 db.make_new_workspace::<String>(&[]); //ID 2
db.update_worktrees(&WorkspaceId(1), &["/tmp", "/tmp2"]); db.update_worktrees(&WorkspaceId(1), &["/tmp", "/tmp2"]);
db.write_to("test.db").unwrap();
// Sanity check // Sanity check
assert_eq!(db.workspace_id(&["/tmp", "/tmp2"]), Some(WorkspaceId(1))); assert_eq!(db.workspace_id(&["/tmp", "/tmp2"]), Some(WorkspaceId(1)));
@ -436,7 +447,7 @@ mod tests {
(WorkspaceId(7), vec!["/tmp2"]), (WorkspaceId(7), vec!["/tmp2"]),
]; ];
let db = Db::open_in_memory(); let db = Db::open_in_memory("test_more_workspace_ids");
for (workspace_id, entries) in data { for (workspace_id, entries) in data {
db.make_new_workspace::<String>(&[]); db.make_new_workspace::<String>(&[]);
@ -470,7 +481,7 @@ mod tests {
(WorkspaceId(3), vec!["/tmp", "/tmp2", "/tmp3"]), (WorkspaceId(3), vec!["/tmp", "/tmp2", "/tmp3"]),
]; ];
let db = Db::open_in_memory(); let db = Db::open_in_memory("test_detect_workspace_id");
for (workspace_id, entries) in data { for (workspace_id, entries) in data {
db.make_new_workspace::<String>(&[]); db.make_new_workspace::<String>(&[]);
@ -511,7 +522,7 @@ mod tests {
(WorkspaceId(3), vec!["/tmp2", "/tmp3"]), (WorkspaceId(3), vec!["/tmp2", "/tmp3"]),
]; ];
let db = Db::open_in_memory(); let db = Db::open_in_memory("test_tricky_overlapping_update");
// Load in the test data // Load in the test data
for (workspace_id, entries) in data { for (workspace_id, entries) in data {
@ -519,6 +530,7 @@ mod tests {
db.update_worktrees(workspace_id, entries); db.update_worktrees(workspace_id, entries);
} }
sleep(Duration::from_secs(1));
// Execute the update // Execute the update
db.update_worktrees(&WorkspaceId(2), &["/tmp2", "/tmp3"]); db.update_worktrees(&WorkspaceId(2), &["/tmp2", "/tmp3"]);

View file

@ -32,6 +32,9 @@ impl Connection {
0 as *const _, 0 as *const _,
); );
// Turn on extended error codes
sqlite3_extended_result_codes(connection.sqlite3, 1);
connection.last_error()?; connection.last_error()?;
} }
@ -71,6 +74,7 @@ impl Connection {
0 as *mut _, 0 as *mut _,
0 as *mut _, 0 as *mut _,
); );
sqlite3_errcode(self.sqlite3);
self.last_error()?; self.last_error()?;
} }
Ok(()) Ok(())
@ -95,29 +99,7 @@ impl Connection {
} }
pub(crate) fn last_error(&self) -> Result<()> { pub(crate) fn last_error(&self) -> Result<()> {
const NON_ERROR_CODES: &[i32] = &[SQLITE_OK, SQLITE_ROW]; unsafe { error_to_result(sqlite3_errcode(self.sqlite3)) }
unsafe {
let code = sqlite3_errcode(self.sqlite3);
if NON_ERROR_CODES.contains(&code) {
return Ok(());
}
let message = sqlite3_errmsg(self.sqlite3);
let message = if message.is_null() {
None
} else {
Some(
String::from_utf8_lossy(CStr::from_ptr(message as *const _).to_bytes())
.into_owned(),
)
};
Err(anyhow!(
"Sqlite call failed with code {} and message: {:?}",
code as isize,
message
))
}
} }
} }
@ -127,12 +109,37 @@ impl Drop for Connection {
} }
} }
pub(crate) fn error_to_result(code: std::os::raw::c_int) -> Result<()> {
const NON_ERROR_CODES: &[i32] = &[SQLITE_OK, SQLITE_ROW];
unsafe {
if NON_ERROR_CODES.contains(&code) {
return Ok(());
}
let message = sqlite3_errstr(code);
let message = if message.is_null() {
None
} else {
Some(
String::from_utf8_lossy(CStr::from_ptr(message as *const _).to_bytes())
.into_owned(),
)
};
Err(anyhow!(
"Sqlite call failed with code {} and message: {:?}",
code as isize,
message
))
}
}
#[cfg(test)] #[cfg(test)]
mod test { mod test {
use anyhow::Result; use anyhow::Result;
use indoc::indoc; use indoc::indoc;
use crate::connection::Connection; use crate::{connection::Connection, migrations::Migration};
#[test] #[test]
fn string_round_trips() -> Result<()> { fn string_round_trips() -> Result<()> {
@ -234,4 +241,34 @@ mod test {
.unwrap(); .unwrap();
assert_eq!(read_blobs, vec![blob]); assert_eq!(read_blobs, vec![blob]);
} }
#[test]
fn test_kv_store() -> anyhow::Result<()> {
let connection = Connection::open_memory("kv_store");
Migration::new(
"kv",
&["CREATE TABLE kv_store(
key TEXT PRIMARY KEY,
value TEXT NOT NULL
) STRICT;"],
)
.run(&connection)
.unwrap();
let mut stmt = connection.prepare("INSERT INTO kv_store(key, value) VALUES(?, ?)")?;
stmt.bind_text(1, "a").unwrap();
stmt.bind_text(2, "b").unwrap();
stmt.exec().unwrap();
let id = connection.last_insert_id();
let res = connection
.prepare("SELECT key, value FROM kv_store WHERE rowid = ?")?
.with_bindings(id)?
.row::<(String, String)>()?;
assert_eq!(res, ("a".to_string(), "b".to_string()));
Ok(())
}
} }

View file

@ -6,9 +6,9 @@ impl Connection {
// Run a set of commands within the context of a `SAVEPOINT name`. If the callback // Run a set of commands within the context of a `SAVEPOINT name`. If the callback
// returns Err(_), the savepoint will be rolled back. Otherwise, the save // returns Err(_), the savepoint will be rolled back. Otherwise, the save
// point is released. // point is released.
pub fn with_savepoint<R, F>(&mut self, name: impl AsRef<str>, f: F) -> Result<R> pub fn with_savepoint<R, F>(&self, name: impl AsRef<str>, f: F) -> Result<R>
where where
F: FnOnce(&mut Connection) -> Result<R>, F: FnOnce(&Connection) -> Result<R>,
{ {
let name = name.as_ref().to_owned(); let name = name.as_ref().to_owned();
self.exec(format!("SAVEPOINT {}", &name))?; self.exec(format!("SAVEPOINT {}", &name))?;
@ -28,13 +28,9 @@ impl Connection {
// Run a set of commands within the context of a `SAVEPOINT name`. If the callback // Run a set of commands within the context of a `SAVEPOINT name`. If the callback
// returns Ok(None) or Err(_), the savepoint will be rolled back. Otherwise, the save // returns Ok(None) or Err(_), the savepoint will be rolled back. Otherwise, the save
// point is released. // point is released.
pub fn with_savepoint_rollback<R, F>( pub fn with_savepoint_rollback<R, F>(&self, name: impl AsRef<str>, f: F) -> Result<Option<R>>
&mut self,
name: impl AsRef<str>,
f: F,
) -> Result<Option<R>>
where where
F: FnOnce(&mut Connection) -> Result<Option<R>>, F: FnOnce(&Connection) -> Result<Option<R>>,
{ {
let name = name.as_ref().to_owned(); let name = name.as_ref().to_owned();
self.exec(format!("SAVEPOINT {}", &name))?; self.exec(format!("SAVEPOINT {}", &name))?;
@ -60,7 +56,7 @@ mod tests {
#[test] #[test]
fn test_nested_savepoints() -> Result<()> { fn test_nested_savepoints() -> Result<()> {
let mut connection = Connection::open_memory("nested_savepoints"); let connection = Connection::open_memory("nested_savepoints");
connection connection
.exec(indoc! {" .exec(indoc! {"

View file

@ -6,7 +6,7 @@ use anyhow::{anyhow, Context, Result};
use libsqlite3_sys::*; use libsqlite3_sys::*;
use crate::bindable::{Bind, Column}; use crate::bindable::{Bind, Column};
use crate::connection::Connection; use crate::connection::{error_to_result, Connection};
pub struct Statement<'a> { pub struct Statement<'a> {
raw_statement: *mut sqlite3_stmt, raw_statement: *mut sqlite3_stmt,
@ -65,6 +65,7 @@ impl<'a> Statement<'a> {
} }
pub fn bind_blob(&self, index: i32, blob: &[u8]) -> Result<()> { pub fn bind_blob(&self, index: i32, blob: &[u8]) -> Result<()> {
// dbg!("bind blob", index);
let index = index as c_int; let index = index as c_int;
let blob_pointer = blob.as_ptr() as *const _; let blob_pointer = blob.as_ptr() as *const _;
let len = blob.len() as c_int; let len = blob.len() as c_int;
@ -94,6 +95,7 @@ impl<'a> Statement<'a> {
} }
pub fn bind_double(&self, index: i32, double: f64) -> Result<()> { pub fn bind_double(&self, index: i32, double: f64) -> Result<()> {
// dbg!("bind double", index);
let index = index as c_int; let index = index as c_int;
unsafe { unsafe {
@ -110,6 +112,7 @@ impl<'a> Statement<'a> {
} }
pub fn bind_int(&self, index: i32, int: i32) -> Result<()> { pub fn bind_int(&self, index: i32, int: i32) -> Result<()> {
// dbg!("bind int", index);
let index = index as c_int; let index = index as c_int;
unsafe { unsafe {
@ -126,6 +129,7 @@ impl<'a> Statement<'a> {
} }
pub fn bind_int64(&self, index: i32, int: i64) -> Result<()> { pub fn bind_int64(&self, index: i32, int: i64) -> Result<()> {
// dbg!("bind int64", index);
let index = index as c_int; let index = index as c_int;
unsafe { unsafe {
sqlite3_bind_int64(self.raw_statement, index, int); sqlite3_bind_int64(self.raw_statement, index, int);
@ -141,6 +145,7 @@ impl<'a> Statement<'a> {
} }
pub fn bind_null(&self, index: i32) -> Result<()> { pub fn bind_null(&self, index: i32) -> Result<()> {
// dbg!("bind null", index);
let index = index as c_int; let index = index as c_int;
unsafe { unsafe {
sqlite3_bind_null(self.raw_statement, index); sqlite3_bind_null(self.raw_statement, index);
@ -149,11 +154,12 @@ impl<'a> Statement<'a> {
} }
pub fn bind_text(&self, index: i32, text: &str) -> Result<()> { pub fn bind_text(&self, index: i32, text: &str) -> Result<()> {
// dbg!("bind text", index, text);
let index = index as c_int; let index = index as c_int;
let text_pointer = text.as_ptr() as *const _; let text_pointer = text.as_ptr() as *const _;
let len = text.len() as c_int; let len = text.len() as c_int;
unsafe { unsafe {
sqlite3_bind_blob( sqlite3_bind_text(
self.raw_statement, self.raw_statement,
index, index,
text_pointer, text_pointer,
@ -304,10 +310,8 @@ impl<'a> Statement<'a> {
impl<'a> Drop for Statement<'a> { impl<'a> Drop for Statement<'a> {
fn drop(&mut self) { fn drop(&mut self) {
unsafe { unsafe {
sqlite3_finalize(self.raw_statement); let error = sqlite3_finalize(self.raw_statement);
self.connection error_to_result(error).expect("failed error");
.last_error()
.expect("sqlite3 finalize failed for statement :(");
}; };
} }
} }

View file

@ -204,6 +204,13 @@ impl<T: Rng> Iterator for RandomCharIter<T> {
} }
} }
#[macro_export]
macro_rules! iife {
($block:block) => {
(|| $block)()
};
}
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use super::*; use super::*;
@ -221,4 +228,18 @@ mod tests {
extend_sorted(&mut vec, vec![1000, 19, 17, 9, 5], 8, |a, b| b.cmp(a)); extend_sorted(&mut vec, vec![1000, 19, 17, 9, 5], 8, |a, b| b.cmp(a));
assert_eq!(vec, &[1000, 101, 21, 19, 17, 13, 9, 8]); assert_eq!(vec, &[1000, 101, 21, 19, 17, 13, 9, 8]);
} }
#[test]
fn test_iife() {
fn option_returning_function() -> Option<()> {
None
}
let foo = iife!({
option_returning_function()?;
Some(())
});
assert_eq!(foo, None);
}
} }