Start work on using sqlite in tests
This commit is contained in:
parent
d14dd27cdc
commit
7e02ac772a
7 changed files with 250 additions and 96 deletions
37
Cargo.lock
generated
37
Cargo.lock
generated
|
@ -1953,6 +1953,18 @@ version = "0.2.0"
|
|||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "7bad48618fdb549078c333a7a8528acb57af271d0433bdecd523eb620628364e"
|
||||
|
||||
[[package]]
|
||||
name = "flume"
|
||||
version = "0.10.14"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "1657b4441c3403d9f7b3409e47575237dac27b1b5726df654a6ecbf92f0f7577"
|
||||
dependencies = [
|
||||
"futures-core",
|
||||
"futures-sink",
|
||||
"pin-project",
|
||||
"spin 0.9.4",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "fnv"
|
||||
version = "1.0.7"
|
||||
|
@ -3022,7 +3034,7 @@ version = "1.4.0"
|
|||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "e2abad23fbc42b3700f2f279844dc832adb2b2eb069b2df918f455c4e18cc646"
|
||||
dependencies = [
|
||||
"spin",
|
||||
"spin 0.5.2",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
|
@ -4725,7 +4737,7 @@ dependencies = [
|
|||
"cc",
|
||||
"libc",
|
||||
"once_cell",
|
||||
"spin",
|
||||
"spin 0.5.2",
|
||||
"untrusted",
|
||||
"web-sys",
|
||||
"winapi 0.3.9",
|
||||
|
@ -5563,6 +5575,15 @@ version = "0.5.2"
|
|||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "6e63cff320ae2c57904679ba7cb63280a3dc4613885beafb148ee7bf9aa9042d"
|
||||
|
||||
[[package]]
|
||||
name = "spin"
|
||||
version = "0.9.4"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "7f6002a767bff9e83f8eeecf883ecb8011875a21ae8da43bffb817a57e78cc09"
|
||||
dependencies = [
|
||||
"lock_api",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "spsc-buffer"
|
||||
version = "0.1.1"
|
||||
|
@ -5583,8 +5604,6 @@ dependencies = [
|
|||
[[package]]
|
||||
name = "sqlx"
|
||||
version = "0.6.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "9249290c05928352f71c077cc44a464d880c63f26f7534728cca008e135c0428"
|
||||
dependencies = [
|
||||
"sqlx-core",
|
||||
"sqlx-macros",
|
||||
|
@ -5593,8 +5612,6 @@ dependencies = [
|
|||
[[package]]
|
||||
name = "sqlx-core"
|
||||
version = "0.6.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "dcbc16ddba161afc99e14d1713a453747a2b07fc097d2009f4c300ec99286105"
|
||||
dependencies = [
|
||||
"ahash",
|
||||
"atoi",
|
||||
|
@ -5608,8 +5625,10 @@ dependencies = [
|
|||
"dotenvy",
|
||||
"either",
|
||||
"event-listener",
|
||||
"flume",
|
||||
"futures-channel",
|
||||
"futures-core",
|
||||
"futures-executor",
|
||||
"futures-intrusive",
|
||||
"futures-util",
|
||||
"hashlink",
|
||||
|
@ -5619,6 +5638,7 @@ dependencies = [
|
|||
"indexmap",
|
||||
"itoa",
|
||||
"libc",
|
||||
"libsqlite3-sys",
|
||||
"log",
|
||||
"md-5",
|
||||
"memchr",
|
||||
|
@ -5648,8 +5668,6 @@ dependencies = [
|
|||
[[package]]
|
||||
name = "sqlx-macros"
|
||||
version = "0.6.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "b850fa514dc11f2ee85be9d055c512aa866746adfacd1cb42d867d68e6a5b0d9"
|
||||
dependencies = [
|
||||
"dotenvy",
|
||||
"either",
|
||||
|
@ -5657,6 +5675,7 @@ dependencies = [
|
|||
"once_cell",
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"serde_json",
|
||||
"sha2 0.10.6",
|
||||
"sqlx-core",
|
||||
"sqlx-rt",
|
||||
|
@ -5667,8 +5686,6 @@ dependencies = [
|
|||
[[package]]
|
||||
name = "sqlx-rt"
|
||||
version = "0.6.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "24c5b2d25fa654cc5f841750b8e1cdedbe21189bf9a9382ee90bfa9dd3562396"
|
||||
dependencies = [
|
||||
"once_cell",
|
||||
"tokio",
|
||||
|
|
|
@ -50,8 +50,9 @@ tracing-log = "0.1.3"
|
|||
tracing-subscriber = { version = "0.3.11", features = ["env-filter", "json"] }
|
||||
|
||||
[dependencies.sqlx]
|
||||
version = "0.6"
|
||||
features = ["runtime-tokio-rustls", "postgres", "time", "uuid"]
|
||||
# version = "0.6"
|
||||
path = "../../../sqlx"
|
||||
features = ["runtime-tokio-rustls", "postgres", "json", "time", "uuid"]
|
||||
|
||||
[dev-dependencies]
|
||||
collections = { path = "../collections", features = ["test-support"] }
|
||||
|
@ -78,5 +79,10 @@ lazy_static = "1.4"
|
|||
serde_json = { version = "1.0", features = ["preserve_order"] }
|
||||
unindent = "0.1"
|
||||
|
||||
[dev-dependencies.sqlx]
|
||||
# version = "0.6"
|
||||
path = "../../../sqlx"
|
||||
features = ["sqlite"]
|
||||
|
||||
[features]
|
||||
seed-support = ["clap", "lipsum", "reqwest"]
|
||||
|
|
127
crates/collab/migrations.sqlite/20221109000000_test_schema.sql
Normal file
127
crates/collab/migrations.sqlite/20221109000000_test_schema.sql
Normal file
|
@ -0,0 +1,127 @@
|
|||
CREATE TABLE IF NOT EXISTS "sessions" (
|
||||
"id" VARCHAR NOT NULL PRIMARY KEY,
|
||||
"expires" TIMESTAMP WITH TIME ZONE NULL,
|
||||
"session" TEXT NOT NULL
|
||||
);
|
||||
|
||||
CREATE TABLE IF NOT EXISTS "users" (
|
||||
"id" INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||
"github_login" VARCHAR,
|
||||
"admin" BOOLEAN,
|
||||
email_address VARCHAR(255) DEFAULT NULL,
|
||||
invite_code VARCHAR(64),
|
||||
invite_count INTEGER NOT NULL DEFAULT 0,
|
||||
inviter_id INTEGER REFERENCES users (id),
|
||||
connected_once BOOLEAN NOT NULL DEFAULT false,
|
||||
created_at TIMESTAMP NOT NULL DEFAULT now,
|
||||
"github_user_id" INTEGER
|
||||
);
|
||||
CREATE UNIQUE INDEX "index_users_github_login" ON "users" ("github_login");
|
||||
CREATE UNIQUE INDEX "index_invite_code_users" ON "users" ("invite_code");
|
||||
CREATE INDEX "index_users_on_email_address" ON "users" ("email_address");
|
||||
CREATE INDEX "index_users_on_github_user_id" ON "users" ("github_user_id");
|
||||
|
||||
CREATE TABLE IF NOT EXISTS "access_tokens" (
|
||||
"id" INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||
"user_id" INTEGER REFERENCES users (id),
|
||||
"hash" VARCHAR(128)
|
||||
);
|
||||
CREATE INDEX "index_access_tokens_user_id" ON "access_tokens" ("user_id");
|
||||
|
||||
CREATE TABLE IF NOT EXISTS "orgs" (
|
||||
"id" SERIAL PRIMARY KEY,
|
||||
"name" VARCHAR NOT NULL,
|
||||
"slug" VARCHAR NOT NULL
|
||||
);
|
||||
CREATE UNIQUE INDEX "index_orgs_slug" ON "orgs" ("slug");
|
||||
|
||||
CREATE TABLE IF NOT EXISTS "org_memberships" (
|
||||
"id" SERIAL PRIMARY KEY,
|
||||
"org_id" INTEGER REFERENCES orgs (id) NOT NULL,
|
||||
"user_id" INTEGER REFERENCES users (id) NOT NULL,
|
||||
"admin" BOOLEAN NOT NULL
|
||||
);
|
||||
CREATE INDEX "index_org_memberships_user_id" ON "org_memberships" ("user_id");
|
||||
CREATE UNIQUE INDEX "index_org_memberships_org_id_and_user_id" ON "org_memberships" ("org_id", "user_id");
|
||||
|
||||
CREATE TABLE IF NOT EXISTS "channels" (
|
||||
"id" SERIAL PRIMARY KEY,
|
||||
"owner_id" INTEGER NOT NULL,
|
||||
"owner_is_user" BOOLEAN NOT NULL,
|
||||
"name" VARCHAR NOT NULL
|
||||
);
|
||||
CREATE UNIQUE INDEX "index_channels_owner_and_name" ON "channels" ("owner_is_user", "owner_id", "name");
|
||||
|
||||
CREATE TABLE IF NOT EXISTS "channel_memberships" (
|
||||
"id" SERIAL PRIMARY KEY,
|
||||
"channel_id" INTEGER REFERENCES channels (id) NOT NULL,
|
||||
"user_id" INTEGER REFERENCES users (id) NOT NULL,
|
||||
"admin" BOOLEAN NOT NULL
|
||||
);
|
||||
CREATE INDEX "index_channel_memberships_user_id" ON "channel_memberships" ("user_id");
|
||||
CREATE UNIQUE INDEX "index_channel_memberships_channel_id_and_user_id" ON "channel_memberships" ("channel_id", "user_id");
|
||||
|
||||
CREATE TABLE IF NOT EXISTS "channel_messages" (
|
||||
"id" SERIAL PRIMARY KEY,
|
||||
"channel_id" INTEGER REFERENCES channels (id) NOT NULL,
|
||||
"sender_id" INTEGER REFERENCES users (id) NOT NULL,
|
||||
"body" TEXT NOT NULL,
|
||||
"sent_at" TIMESTAMP
|
||||
);
|
||||
CREATE INDEX "index_channel_messages_channel_id" ON "channel_messages" ("channel_id");
|
||||
|
||||
CREATE TABLE IF NOT EXISTS "contacts" (
|
||||
"id" SERIAL PRIMARY KEY,
|
||||
"user_id_a" INTEGER REFERENCES users (id) NOT NULL,
|
||||
"user_id_b" INTEGER REFERENCES users (id) NOT NULL,
|
||||
"a_to_b" BOOLEAN NOT NULL,
|
||||
"should_notify" BOOLEAN NOT NULL,
|
||||
"accepted" BOOLEAN NOT NULL
|
||||
);
|
||||
CREATE UNIQUE INDEX "index_contacts_user_ids" ON "contacts" ("user_id_a", "user_id_b");
|
||||
CREATE INDEX "index_contacts_user_id_b" ON "contacts" ("user_id_b");
|
||||
|
||||
CREATE TABLE IF NOT EXISTS "projects" (
|
||||
"id" SERIAL PRIMARY KEY,
|
||||
"host_user_id" INTEGER REFERENCES users (id) NOT NULL,
|
||||
"unregistered" BOOLEAN NOT NULL DEFAULT false
|
||||
);
|
||||
|
||||
CREATE TABLE IF NOT EXISTS "worktree_extensions" (
|
||||
"id" SERIAL PRIMARY KEY,
|
||||
"project_id" INTEGER REFERENCES projects (id) NOT NULL,
|
||||
"worktree_id" INTEGER NOT NULL,
|
||||
"extension" VARCHAR(255),
|
||||
"count" INTEGER NOT NULL
|
||||
);
|
||||
CREATE UNIQUE INDEX "index_worktree_extensions_on_project_id_and_worktree_id_and_extension" ON "worktree_extensions" ("project_id", "worktree_id", "extension");
|
||||
|
||||
CREATE TABLE IF NOT EXISTS "project_activity_periods" (
|
||||
"id" SERIAL PRIMARY KEY,
|
||||
"duration_millis" INTEGER NOT NULL,
|
||||
"ended_at" TIMESTAMP NOT NULL,
|
||||
"user_id" INTEGER REFERENCES users (id) NOT NULL,
|
||||
"project_id" INTEGER REFERENCES projects (id) NOT NULL
|
||||
);
|
||||
CREATE INDEX "index_project_activity_periods_on_ended_at" ON "project_activity_periods" ("ended_at");
|
||||
|
||||
CREATE TABLE IF NOT EXISTS "signups" (
|
||||
"id" SERIAL PRIMARY KEY,
|
||||
"email_address" VARCHAR NOT NULL,
|
||||
"email_confirmation_code" VARCHAR(64) NOT NULL,
|
||||
"email_confirmation_sent" BOOLEAN NOT NULL,
|
||||
"created_at" TIMESTAMP NOT NULL DEFAULT CURRENT_TIMESTAMP,
|
||||
"device_id" VARCHAR,
|
||||
"user_id" INTEGER REFERENCES users (id) ON DELETE CASCADE,
|
||||
"inviting_user_id" INTEGER REFERENCES users (id) ON DELETE SET NULL,
|
||||
|
||||
"platform_mac" BOOLEAN NOT NULL,
|
||||
"platform_linux" BOOLEAN NOT NULL,
|
||||
"platform_windows" BOOLEAN NOT NULL,
|
||||
"platform_unknown" BOOLEAN NOT NULL,
|
||||
|
||||
"editor_features" VARCHAR[],
|
||||
"programming_languages" VARCHAR[]
|
||||
);
|
||||
CREATE UNIQUE INDEX "index_signups_on_email_address" ON "signups" ("email_address");
|
||||
CREATE INDEX "index_signups_on_email_confirmation_sent" ON "signups" ("email_confirmation_sent");
|
|
@ -5,7 +5,6 @@ use axum::http::StatusCode;
|
|||
use collections::HashMap;
|
||||
use futures::StreamExt;
|
||||
use serde::{Deserialize, Serialize};
|
||||
pub use sqlx::postgres::PgPoolOptions as DbOptions;
|
||||
use sqlx::{
|
||||
migrate::{Migrate as _, Migration, MigrationSource},
|
||||
types::Uuid,
|
||||
|
@ -181,11 +180,14 @@ pub trait Db: Send + Sync {
|
|||
pub const DEFAULT_MIGRATIONS_PATH: Option<&'static str> =
|
||||
Some(concat!(env!("CARGO_MANIFEST_DIR"), "/migrations"));
|
||||
|
||||
pub const TEST_MIGRATIONS_PATH: Option<&'static str> =
|
||||
Some(concat!(env!("CARGO_MANIFEST_DIR"), "/migrations.sqlite"));
|
||||
|
||||
#[cfg(not(any(test, debug_assertions)))]
|
||||
pub const DEFAULT_MIGRATIONS_PATH: Option<&'static str> = None;
|
||||
|
||||
pub struct PostgresDb {
|
||||
pool: sqlx::PgPool,
|
||||
pub struct RealDb {
|
||||
pool: sqlx::SqlitePool,
|
||||
}
|
||||
|
||||
macro_rules! test_support {
|
||||
|
@ -202,13 +204,13 @@ macro_rules! test_support {
|
|||
}};
|
||||
}
|
||||
|
||||
impl PostgresDb {
|
||||
impl RealDb {
|
||||
pub async fn new(url: &str, max_connections: u32) -> Result<Self> {
|
||||
let pool = DbOptions::new()
|
||||
.max_connections(max_connections)
|
||||
eprintln!("{url}");
|
||||
let pool = sqlx::sqlite::SqlitePoolOptions::new()
|
||||
.max_connections(1)
|
||||
.connect(url)
|
||||
.await
|
||||
.context("failed to connect to postgres database")?;
|
||||
.await?;
|
||||
Ok(Self { pool })
|
||||
}
|
||||
|
||||
|
@ -267,7 +269,7 @@ impl PostgresDb {
|
|||
}
|
||||
|
||||
#[async_trait]
|
||||
impl Db for PostgresDb {
|
||||
impl Db for RealDb {
|
||||
// users
|
||||
|
||||
async fn create_user(
|
||||
|
@ -280,8 +282,8 @@ impl Db for PostgresDb {
|
|||
let query = "
|
||||
INSERT INTO users (email_address, github_login, github_user_id, admin)
|
||||
VALUES ($1, $2, $3, $4)
|
||||
ON CONFLICT (github_login) DO UPDATE SET github_login = excluded.github_login
|
||||
RETURNING id, metrics_id::text
|
||||
-- ON CONFLICT (github_login) DO UPDATE SET github_login = excluded.github_login
|
||||
RETURNING id, 'the-metrics-id'
|
||||
";
|
||||
|
||||
let (user_id, metrics_id): (UserId, String) = sqlx::query_as(query)
|
||||
|
@ -331,8 +333,18 @@ impl Db for PostgresDb {
|
|||
}
|
||||
|
||||
async fn get_user_by_id(&self, id: UserId) -> Result<Option<User>> {
|
||||
let users = self.get_users_by_ids(vec![id]).await?;
|
||||
Ok(users.into_iter().next())
|
||||
test_support!(self, {
|
||||
let query = "
|
||||
SELECT users.*
|
||||
FROM users
|
||||
WHERE id = $1
|
||||
LIMIT 1
|
||||
";
|
||||
Ok(sqlx::query_as(query)
|
||||
.bind(&id)
|
||||
.fetch_optional(&self.pool)
|
||||
.await?)
|
||||
})
|
||||
}
|
||||
|
||||
async fn get_user_metrics_id(&self, id: UserId) -> Result<String> {
|
||||
|
@ -351,14 +363,13 @@ impl Db for PostgresDb {
|
|||
|
||||
async fn get_users_by_ids(&self, ids: Vec<UserId>) -> Result<Vec<User>> {
|
||||
test_support!(self, {
|
||||
let ids = ids.into_iter().map(|id| id.0).collect::<Vec<_>>();
|
||||
let query = "
|
||||
SELECT users.*
|
||||
FROM users
|
||||
WHERE users.id = ANY ($1)
|
||||
WHERE users.id IN (SELECT value from json_each($1))
|
||||
";
|
||||
Ok(sqlx::query_as(query)
|
||||
.bind(&ids)
|
||||
.bind(&serde_json::json!(ids))
|
||||
.fetch_all(&self.pool)
|
||||
.await?)
|
||||
})
|
||||
|
@ -493,7 +504,7 @@ impl Db for PostgresDb {
|
|||
device_id
|
||||
)
|
||||
VALUES
|
||||
($1, $2, 'f', $3, $4, $5, 'f', $6, $7, $8)
|
||||
($1, $2, FALSE, $3, $4, $5, FALSE, $6)
|
||||
RETURNING id
|
||||
",
|
||||
)
|
||||
|
@ -502,8 +513,8 @@ impl Db for PostgresDb {
|
|||
.bind(&signup.platform_linux)
|
||||
.bind(&signup.platform_mac)
|
||||
.bind(&signup.platform_windows)
|
||||
.bind(&signup.editor_features)
|
||||
.bind(&signup.programming_languages)
|
||||
// .bind(&signup.editor_features)
|
||||
// .bind(&signup.programming_languages)
|
||||
.bind(&signup.device_id)
|
||||
.execute(&self.pool)
|
||||
.await?;
|
||||
|
@ -555,21 +566,21 @@ impl Db for PostgresDb {
|
|||
|
||||
async fn record_sent_invites(&self, invites: &[Invite]) -> Result<()> {
|
||||
test_support!(self, {
|
||||
sqlx::query(
|
||||
"
|
||||
UPDATE signups
|
||||
SET email_confirmation_sent = 't'
|
||||
WHERE email_address = ANY ($1)
|
||||
",
|
||||
)
|
||||
.bind(
|
||||
&invites
|
||||
.iter()
|
||||
.map(|s| s.email_address.as_str())
|
||||
.collect::<Vec<_>>(),
|
||||
)
|
||||
.execute(&self.pool)
|
||||
.await?;
|
||||
// sqlx::query(
|
||||
// "
|
||||
// UPDATE signups
|
||||
// SET email_confirmation_sent = TRUE
|
||||
// WHERE email_address = ANY ($1)
|
||||
// ",
|
||||
// )
|
||||
// .bind(
|
||||
// &invites
|
||||
// .iter()
|
||||
// .map(|s| s.email_address.as_str())
|
||||
// .collect::<Vec<_>>(),
|
||||
// )
|
||||
// .execute(&self.pool)
|
||||
// .await?;
|
||||
Ok(())
|
||||
})
|
||||
}
|
||||
|
@ -611,7 +622,7 @@ impl Db for PostgresDb {
|
|||
INSERT INTO users
|
||||
(email_address, github_login, github_user_id, admin, invite_count, invite_code)
|
||||
VALUES
|
||||
($1, $2, $3, 'f', $4, $5)
|
||||
($1, $2, $3, FALSE, $4, $5)
|
||||
ON CONFLICT (github_login) DO UPDATE SET
|
||||
email_address = excluded.email_address,
|
||||
github_user_id = excluded.github_user_id,
|
||||
|
@ -664,7 +675,7 @@ impl Db for PostgresDb {
|
|||
INSERT INTO contacts
|
||||
(user_id_a, user_id_b, a_to_b, should_notify, accepted)
|
||||
VALUES
|
||||
($1, $2, 't', 't', 't')
|
||||
($1, $2, TRUE, TRUE, TRUE)
|
||||
ON CONFLICT DO NOTHING
|
||||
",
|
||||
)
|
||||
|
@ -824,7 +835,7 @@ impl Db for PostgresDb {
|
|||
device_id
|
||||
)
|
||||
VALUES
|
||||
($1, $2, 'f', $3, 'f', 'f', 'f', 't', $4)
|
||||
($1, $2, FALSE, $3, FALSE, FALSE, FALSE, TRUE, $4)
|
||||
ON CONFLICT (email_address)
|
||||
DO UPDATE SET
|
||||
inviting_user_id = excluded.inviting_user_id
|
||||
|
@ -870,7 +881,7 @@ impl Db for PostgresDb {
|
|||
sqlx::query(
|
||||
"
|
||||
UPDATE projects
|
||||
SET unregistered = 't'
|
||||
SET unregistered = TRUE
|
||||
WHERE id = $1
|
||||
",
|
||||
)
|
||||
|
@ -1274,7 +1285,7 @@ impl Db for PostgresDb {
|
|||
|
||||
let query = "
|
||||
SELECT 1 FROM contacts
|
||||
WHERE user_id_a = $1 AND user_id_b = $2 AND accepted = 't'
|
||||
WHERE user_id_a = $1 AND user_id_b = $2 AND accepted = TRUE
|
||||
LIMIT 1
|
||||
";
|
||||
Ok(sqlx::query_scalar::<_, i32>(query)
|
||||
|
@ -1295,11 +1306,11 @@ impl Db for PostgresDb {
|
|||
};
|
||||
let query = "
|
||||
INSERT into contacts (user_id_a, user_id_b, a_to_b, accepted, should_notify)
|
||||
VALUES ($1, $2, $3, 'f', 't')
|
||||
VALUES ($1, $2, $3, FALSE, TRUE)
|
||||
ON CONFLICT (user_id_a, user_id_b) DO UPDATE
|
||||
SET
|
||||
accepted = 't',
|
||||
should_notify = 'f'
|
||||
accepted = TRUE,
|
||||
should_notify = FALSE
|
||||
WHERE
|
||||
NOT contacts.accepted AND
|
||||
((contacts.a_to_b = excluded.a_to_b AND contacts.user_id_a = excluded.user_id_b) OR
|
||||
|
@ -1359,7 +1370,7 @@ impl Db for PostgresDb {
|
|||
|
||||
let query = "
|
||||
UPDATE contacts
|
||||
SET should_notify = 'f'
|
||||
SET should_notify = FALSE
|
||||
WHERE
|
||||
user_id_a = $1 AND user_id_b = $2 AND
|
||||
(
|
||||
|
@ -1398,7 +1409,7 @@ impl Db for PostgresDb {
|
|||
let result = if accept {
|
||||
let query = "
|
||||
UPDATE contacts
|
||||
SET accepted = 't', should_notify = 't'
|
||||
SET accepted = TRUE, should_notify = TRUE
|
||||
WHERE user_id_a = $1 AND user_id_b = $2 AND a_to_b = $3;
|
||||
";
|
||||
sqlx::query(query)
|
||||
|
@ -1706,7 +1717,7 @@ impl Db for PostgresDb {
|
|||
";
|
||||
sqlx::query(query).execute(&self.pool).await.log_err();
|
||||
self.pool.close().await;
|
||||
<sqlx::Postgres as sqlx::migrate::MigrateDatabase>::drop_database(url)
|
||||
<sqlx::Sqlite as sqlx::migrate::MigrateDatabase>::drop_database(url)
|
||||
.await
|
||||
.log_err();
|
||||
eprintln!("tore down database: {:?}", start.elapsed());
|
||||
|
@ -1929,10 +1940,9 @@ mod test {
|
|||
use anyhow::anyhow;
|
||||
use collections::BTreeMap;
|
||||
use gpui::executor::Background;
|
||||
use lazy_static::lazy_static;
|
||||
use parking_lot::Mutex;
|
||||
use rand::prelude::*;
|
||||
use sqlx::{migrate::MigrateDatabase, Postgres};
|
||||
use sqlx::{migrate::MigrateDatabase, Sqlite};
|
||||
use std::sync::Arc;
|
||||
use util::post_inc;
|
||||
|
||||
|
@ -2587,22 +2597,14 @@ mod test {
|
|||
|
||||
impl TestDb {
|
||||
#[allow(clippy::await_holding_lock)]
|
||||
pub async fn postgres() -> Self {
|
||||
lazy_static! {
|
||||
static ref LOCK: Mutex<()> = Mutex::new(());
|
||||
}
|
||||
|
||||
pub async fn real() -> Self {
|
||||
eprintln!("creating database...");
|
||||
let start = std::time::Instant::now();
|
||||
let _guard = LOCK.lock();
|
||||
let mut rng = StdRng::from_entropy();
|
||||
let name = format!("zed-test-{}", rng.gen::<u128>());
|
||||
let url = format!("postgres://postgres@localhost:5433/{}", name);
|
||||
Postgres::create_database(&url)
|
||||
.await
|
||||
.expect("failed to create test db");
|
||||
let db = PostgresDb::new(&url, 5).await.unwrap();
|
||||
db.migrate(Path::new(DEFAULT_MIGRATIONS_PATH.unwrap()), false)
|
||||
let url = format!("/tmp/zed-test-{}", rng.gen::<u128>());
|
||||
Sqlite::create_database(&url).await.unwrap();
|
||||
let db = RealDb::new(&url, 5).await.unwrap();
|
||||
db.migrate(Path::new(TEST_MIGRATIONS_PATH.unwrap()), false)
|
||||
.await
|
||||
.unwrap();
|
||||
|
||||
|
@ -2628,7 +2630,7 @@ mod test {
|
|||
impl Drop for TestDb {
|
||||
fn drop(&mut self) {
|
||||
if let Some(db) = self.db.take() {
|
||||
futures::executor::block_on(db.teardown(&self.url));
|
||||
std::fs::remove_file(&self.url).ok();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -7,7 +7,7 @@ use time::OffsetDateTime;
|
|||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_get_users_by_ids() {
|
||||
for test_db in [
|
||||
TestDb::postgres().await,
|
||||
TestDb::real().await,
|
||||
TestDb::fake(build_background_executor()),
|
||||
] {
|
||||
let db = test_db.db();
|
||||
|
@ -73,7 +73,7 @@ async fn test_get_users_by_ids() {
|
|||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_get_user_by_github_account() {
|
||||
for test_db in [
|
||||
TestDb::postgres().await,
|
||||
TestDb::real().await,
|
||||
TestDb::fake(build_background_executor()),
|
||||
] {
|
||||
let db = test_db.db();
|
||||
|
@ -132,7 +132,7 @@ async fn test_get_user_by_github_account() {
|
|||
|
||||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_worktree_extensions() {
|
||||
let test_db = TestDb::postgres().await;
|
||||
let test_db = TestDb::real().await;
|
||||
let db = test_db.db();
|
||||
|
||||
let user = db
|
||||
|
@ -204,7 +204,7 @@ async fn test_worktree_extensions() {
|
|||
|
||||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_user_activity() {
|
||||
let test_db = TestDb::postgres().await;
|
||||
let test_db = TestDb::real().await;
|
||||
let db = test_db.db();
|
||||
|
||||
let mut user_ids = Vec::new();
|
||||
|
@ -448,7 +448,7 @@ async fn test_user_activity() {
|
|||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_recent_channel_messages() {
|
||||
for test_db in [
|
||||
TestDb::postgres().await,
|
||||
TestDb::real().await,
|
||||
TestDb::fake(build_background_executor()),
|
||||
] {
|
||||
let db = test_db.db();
|
||||
|
@ -493,7 +493,7 @@ async fn test_recent_channel_messages() {
|
|||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_channel_message_nonces() {
|
||||
for test_db in [
|
||||
TestDb::postgres().await,
|
||||
TestDb::real().await,
|
||||
TestDb::fake(build_background_executor()),
|
||||
] {
|
||||
let db = test_db.db();
|
||||
|
@ -538,7 +538,7 @@ async fn test_channel_message_nonces() {
|
|||
|
||||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_create_access_tokens() {
|
||||
let test_db = TestDb::postgres().await;
|
||||
let test_db = TestDb::real().await;
|
||||
let db = test_db.db();
|
||||
let user = db
|
||||
.create_user(
|
||||
|
@ -582,14 +582,14 @@ async fn test_create_access_tokens() {
|
|||
|
||||
#[test]
|
||||
fn test_fuzzy_like_string() {
|
||||
assert_eq!(PostgresDb::fuzzy_like_string("abcd"), "%a%b%c%d%");
|
||||
assert_eq!(PostgresDb::fuzzy_like_string("x y"), "%x%y%");
|
||||
assert_eq!(PostgresDb::fuzzy_like_string(" z "), "%z%");
|
||||
assert_eq!(RealDb::fuzzy_like_string("abcd"), "%a%b%c%d%");
|
||||
assert_eq!(RealDb::fuzzy_like_string("x y"), "%x%y%");
|
||||
assert_eq!(RealDb::fuzzy_like_string(" z "), "%z%");
|
||||
}
|
||||
|
||||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_fuzzy_search_users() {
|
||||
let test_db = TestDb::postgres().await;
|
||||
let test_db = TestDb::real().await;
|
||||
let db = test_db.db();
|
||||
for (i, github_login) in [
|
||||
"California",
|
||||
|
@ -638,7 +638,7 @@ async fn test_fuzzy_search_users() {
|
|||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_add_contacts() {
|
||||
for test_db in [
|
||||
TestDb::postgres().await,
|
||||
TestDb::real().await,
|
||||
TestDb::fake(build_background_executor()),
|
||||
] {
|
||||
let db = test_db.db();
|
||||
|
@ -805,7 +805,7 @@ async fn test_add_contacts() {
|
|||
|
||||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_invite_codes() {
|
||||
let postgres = TestDb::postgres().await;
|
||||
let postgres = TestDb::real().await;
|
||||
let db = postgres.db();
|
||||
let NewUserResult { user_id: user1, .. } = db
|
||||
.create_user(
|
||||
|
@ -1000,7 +1000,7 @@ async fn test_invite_codes() {
|
|||
|
||||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_signups() {
|
||||
let postgres = TestDb::postgres().await;
|
||||
let postgres = TestDb::real().await;
|
||||
let db = postgres.db();
|
||||
|
||||
// people sign up on the waitlist
|
||||
|
@ -1146,7 +1146,7 @@ async fn test_signups() {
|
|||
|
||||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_metrics_id() {
|
||||
let postgres = TestDb::postgres().await;
|
||||
let postgres = TestDb::real().await;
|
||||
let db = postgres.db();
|
||||
|
||||
let NewUserResult {
|
||||
|
|
|
@ -53,7 +53,6 @@ use std::{
|
|||
time::Duration,
|
||||
};
|
||||
use theme::ThemeRegistry;
|
||||
use tokio::runtime::{EnterGuard, Runtime};
|
||||
use unindent::Unindent as _;
|
||||
use util::post_inc;
|
||||
use workspace::{shared_screen::SharedScreen, Item, SplitDirection, ToggleFollow, Workspace};
|
||||
|
@ -80,7 +79,6 @@ async fn test_basic_calls(
|
|||
let mut server = TestServer::start(cx_a.foreground(), cx_a.background()).await;
|
||||
|
||||
let start = std::time::Instant::now();
|
||||
eprintln!("test_basic_calls");
|
||||
|
||||
let client_a = server.create_client(cx_a, "user_a").await;
|
||||
let client_b = server.create_client(cx_b, "user_b").await;
|
||||
|
@ -6106,7 +6104,7 @@ impl TestServer {
|
|||
.enable_time()
|
||||
.build()
|
||||
.unwrap()
|
||||
.block_on(TestDb::postgres());
|
||||
.block_on(TestDb::real());
|
||||
let live_kit_server_id = NEXT_LIVE_KIT_SERVER_ID.fetch_add(1, SeqCst);
|
||||
let live_kit_server = live_kit_client::TestServer::create(
|
||||
format!("http://livekit.{}.test", live_kit_server_id),
|
||||
|
@ -6162,7 +6160,7 @@ impl TestServer {
|
|||
},
|
||||
)
|
||||
.await
|
||||
.unwrap()
|
||||
.expect("creating user failed")
|
||||
.user_id
|
||||
};
|
||||
let client_name = name.to_string();
|
||||
|
@ -6202,7 +6200,11 @@ impl TestServer {
|
|||
let (client_conn, server_conn, killed) =
|
||||
Connection::in_memory(cx.background());
|
||||
let (connection_id_tx, connection_id_rx) = oneshot::channel();
|
||||
let user = db.get_user_by_id(user_id).await.unwrap().unwrap();
|
||||
let user = db
|
||||
.get_user_by_id(user_id)
|
||||
.await
|
||||
.expect("retrieving user failed")
|
||||
.unwrap();
|
||||
cx.background()
|
||||
.spawn(server.handle_connection(
|
||||
server_conn,
|
||||
|
|
|
@ -13,7 +13,7 @@ use crate::rpc::ResultExt as _;
|
|||
use anyhow::anyhow;
|
||||
use axum::{routing::get, Router};
|
||||
use collab::{Error, Result};
|
||||
use db::{Db, PostgresDb};
|
||||
use db::{Db, RealDb};
|
||||
use serde::Deserialize;
|
||||
use std::{
|
||||
env::args,
|
||||
|
@ -56,7 +56,7 @@ pub struct AppState {
|
|||
|
||||
impl AppState {
|
||||
async fn new(config: Config) -> Result<Arc<Self>> {
|
||||
let db = PostgresDb::new(&config.database_url, 5).await?;
|
||||
let db = RealDb::new(&config.database_url, 5).await?;
|
||||
let live_kit_client = if let Some(((server, key), secret)) = config
|
||||
.live_kit_server
|
||||
.as_ref()
|
||||
|
@ -96,7 +96,7 @@ async fn main() -> Result<()> {
|
|||
}
|
||||
Some("migrate") => {
|
||||
let config = envy::from_env::<MigrateConfig>().expect("error loading config");
|
||||
let db = PostgresDb::new(&config.database_url, 5).await?;
|
||||
let db = RealDb::new(&config.database_url, 5).await?;
|
||||
|
||||
let migrations_path = config
|
||||
.migrations_path
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue