parent
4199a97a19
commit
45a3bf291b
7 changed files with 176 additions and 160 deletions
|
|
@ -1,4 +1,4 @@
|
||||||
use lib_sync_core::task_manager::TaskPayload;
|
use lib_sync_core::tasks::TaskPayload;
|
||||||
use chrono::{DateTime, Local};
|
use chrono::{DateTime, Local};
|
||||||
use serde::{de, Deserialize, Deserializer, Serialize};
|
use serde::{de, Deserialize, Deserializer, Serialize};
|
||||||
use serde_json::Value;
|
use serde_json::Value;
|
||||||
|
|
|
||||||
|
|
@ -4,7 +4,7 @@ use figment::{
|
||||||
Figment,
|
Figment,
|
||||||
providers::{Env, Serialized},
|
providers::{Env, Serialized},
|
||||||
};
|
};
|
||||||
use lib_sync_core::task_manager::{TaskManager, TaskStatus};
|
use lib_sync_core::tasks::{TaskStatus};
|
||||||
use cli::config::{Command, Config};
|
use cli::config::{Command, Config};
|
||||||
use cli::{Error, Result};
|
use cli::{Error, Result};
|
||||||
use std::fs::File;
|
use std::fs::File;
|
||||||
|
|
|
||||||
|
|
@ -1,14 +1,6 @@
|
||||||
use crate::task_manager::{Task, TaskPayload, TaskStatus};
|
use crate::tasks::{Task, TaskPayload, TaskStatus};
|
||||||
use futures::stream::BoxStream;
|
use futures::{Stream};
|
||||||
use futures::{Stream, StreamExt, TryStreamExt};
|
mod sqlite;
|
||||||
use serde::Serialize;
|
|
||||||
use sqlx::sqlite::{SqliteConnectOptions, SqliteJournalMode};
|
|
||||||
use sqlx::{Error, QueryBuilder, Sqlite, SqlitePool};
|
|
||||||
use std::path::PathBuf;
|
|
||||||
use tokio::fs;
|
|
||||||
use tracing::{info, instrument};
|
|
||||||
|
|
||||||
static SQLITE_BIND_LIMIT: usize = 32766;
|
|
||||||
|
|
||||||
#[derive(Default, Clone)]
|
#[derive(Default, Clone)]
|
||||||
pub struct TaskPagination {
|
pub struct TaskPagination {
|
||||||
|
|
@ -77,121 +69,9 @@ impl<T: TaskPayload> TasksPage<T> {
|
||||||
}
|
}
|
||||||
|
|
||||||
pub trait TaskStorage<T: TaskPayload> {
|
pub trait TaskStorage<T: TaskPayload> {
|
||||||
fn insert_tasks(&self, tasks: Vec<Task<T>>) -> crate::Result<()>;
|
async fn insert_tasks(&self, tasks: Vec<Task<T>>) -> crate::Result<()>;
|
||||||
fn get_tasks(&self, options: TaskStatus) -> impl Stream<Item = crate::Result<Task<T>>>;
|
fn get_tasks(&self, options: TaskStatus) -> impl Stream<Item = crate::Result<Task<T>>>;
|
||||||
|
|
||||||
async fn get_paginated_tasks(&self, page: &TaskPagination) -> crate::Result<TasksPage<T>>;
|
async fn get_paginated_tasks(&self, page: &TaskPagination) -> crate::Result<TasksPage<T>>;
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug)]
|
|
||||||
pub struct Database {
|
|
||||||
pool: SqlitePool,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Database {
|
|
||||||
pub async fn new<P: Into<PathBuf>>(base_path: P) -> crate::Result<Self> {
|
|
||||||
Ok(Self {
|
|
||||||
pool: Self::connect_database(base_path).await?,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn connect_database<P: Into<PathBuf>>(base_path: P) -> crate::Result<SqlitePool> {
|
|
||||||
let base_path = base_path.into();
|
|
||||||
|
|
||||||
let database_file_path = base_path.join("db.sql");
|
|
||||||
|
|
||||||
fs::create_dir_all(base_path).await?;
|
|
||||||
|
|
||||||
let opts = SqliteConnectOptions::new()
|
|
||||||
.filename(database_file_path)
|
|
||||||
.create_if_missing(true)
|
|
||||||
.journal_mode(SqliteJournalMode::Wal);
|
|
||||||
|
|
||||||
let pool = SqlitePool::connect_with(opts).await?;
|
|
||||||
|
|
||||||
sqlx::migrate!("../migrations").run(&pool).await?;
|
|
||||||
|
|
||||||
Ok(pool)
|
|
||||||
}
|
|
||||||
|
|
||||||
#[instrument(skip(self, values))]
|
|
||||||
pub async fn load_tasks<T>(&self, values: Vec<T>) -> crate::Result<()>
|
|
||||||
where
|
|
||||||
T: TaskPayload + Serialize + std::fmt::Debug,
|
|
||||||
{
|
|
||||||
let mut tx = self.pool.begin().await?;
|
|
||||||
let mut builder: QueryBuilder<'_, Sqlite> =
|
|
||||||
QueryBuilder::new("insert into tasks(payload_key, payload, status_id)");
|
|
||||||
|
|
||||||
let args: crate::Result<Vec<(String, String)>> = values
|
|
||||||
.iter()
|
|
||||||
.map(|value| Ok((value.get_key(), serde_json::to_string(value)?)))
|
|
||||||
.collect();
|
|
||||||
|
|
||||||
let mut affected_rows = 0;
|
|
||||||
// Chunk the query by the size limit of bind params
|
|
||||||
for chunk in args?.chunks(SQLITE_BIND_LIMIT / 3) {
|
|
||||||
builder.push_values(chunk, |mut builder, item| {
|
|
||||||
builder
|
|
||||||
.push_bind(&item.0)
|
|
||||||
.push_bind(&item.1)
|
|
||||||
.push_bind(TaskStatus::Pending);
|
|
||||||
});
|
|
||||||
builder.push("ON conflict (payload_key) DO NOTHING");
|
|
||||||
|
|
||||||
let query = builder.build();
|
|
||||||
|
|
||||||
affected_rows += query.execute(&mut *tx).await?.rows_affected();
|
|
||||||
builder.reset();
|
|
||||||
}
|
|
||||||
|
|
||||||
tx.commit().await?;
|
|
||||||
|
|
||||||
info!("{} rows inserted.", affected_rows);
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<T: TaskPayload> TaskStorage<T> for Database {
|
|
||||||
fn insert_tasks(&self, tasks: Vec<Task<T>>) -> crate::error::Result<()> {
|
|
||||||
todo!()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn get_tasks(&self, task_status: TaskStatus) -> impl Stream<Item = crate::Result<Task<T>>> {
|
|
||||||
let query= sqlx::query_as::<_, Task<T>>(
|
|
||||||
"
|
|
||||||
SELECT id, payload_key, payload, status_id, created_at, updated_at
|
|
||||||
FROM tasks
|
|
||||||
WHERE status_id = ?
|
|
||||||
ORDER BY created_at DESC
|
|
||||||
",
|
|
||||||
).bind(task_status);
|
|
||||||
|
|
||||||
query.fetch(&self.pool).err_into::<crate::Error>()
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn get_paginated_tasks(&self, page: &TaskPagination) -> crate::Result<TasksPage<T>> {
|
|
||||||
let mut builder: QueryBuilder<'_, Sqlite> = QueryBuilder::new(
|
|
||||||
"select id, payload_key, payload, status_id, created_at, updated_at from tasks ",
|
|
||||||
);
|
|
||||||
|
|
||||||
if let Some(status) = &page.status {
|
|
||||||
builder.push("where status_id = ").push_bind(status);
|
|
||||||
}
|
|
||||||
|
|
||||||
builder.push("ORDER BY created_at DESC ");
|
|
||||||
|
|
||||||
if let Some(limit) = &page.offset {
|
|
||||||
builder.push("OFFSET ").push_bind(limit);
|
|
||||||
}
|
|
||||||
|
|
||||||
if let Some(limit) = &page.limit {
|
|
||||||
builder.push("LIMIT ").push_bind(limit);
|
|
||||||
}
|
|
||||||
|
|
||||||
let tasks = builder.build_query_as::<Task<T>>().fetch_all(&self.pool).await?;
|
|
||||||
|
|
||||||
Ok(TasksPage::new(tasks, page.clone()))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
|
||||||
117
lib_sync_core/src/database/sqlite.rs
Normal file
117
lib_sync_core/src/database/sqlite.rs
Normal file
|
|
@ -0,0 +1,117 @@
|
||||||
|
use sqlx::{QueryBuilder, SqlitePool};
|
||||||
|
use std::path::PathBuf;
|
||||||
|
use tokio::fs;
|
||||||
|
use sqlx::sqlite::{SqliteConnectOptions, SqliteJournalMode};
|
||||||
|
use tracing::{info, instrument};
|
||||||
|
use futures::{Stream, TryStreamExt};
|
||||||
|
use crate::database::{TaskPagination, TaskStorage, TasksPage};
|
||||||
|
use crate::tasks::{Task, TaskPayload, TaskStatus};
|
||||||
|
|
||||||
|
#[allow(unused)]
|
||||||
|
static SQLITE_BIND_LIMIT: usize = 32766;
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct Sqlite {
|
||||||
|
pool: SqlitePool,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Sqlite {
|
||||||
|
pub async fn new<P: Into<PathBuf>>(base_path: P) -> crate::Result<Self> {
|
||||||
|
Ok(Self {
|
||||||
|
pool: Self::connect_database(base_path).await?,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn connect_database<P: Into<PathBuf>>(base_path: P) -> crate::Result<SqlitePool> {
|
||||||
|
let base_path = base_path.into();
|
||||||
|
|
||||||
|
let database_file_path = base_path.join("db.sql");
|
||||||
|
|
||||||
|
fs::create_dir_all(base_path).await?;
|
||||||
|
|
||||||
|
let opts = SqliteConnectOptions::new()
|
||||||
|
.filename(database_file_path)
|
||||||
|
.create_if_missing(true)
|
||||||
|
.journal_mode(SqliteJournalMode::Wal);
|
||||||
|
|
||||||
|
let pool = SqlitePool::connect_with(opts).await?;
|
||||||
|
|
||||||
|
sqlx::migrate!("../migrations").run(&pool).await?;
|
||||||
|
|
||||||
|
Ok(pool)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T: TaskPayload> TaskStorage<T> for Sqlite {
|
||||||
|
#[instrument(skip(self, tasks))]
|
||||||
|
async fn insert_tasks(&self, tasks: Vec<Task<T>>) -> crate::Result<()> {
|
||||||
|
let mut tx = self.pool.begin().await?;
|
||||||
|
let mut builder: QueryBuilder<'_, sqlx::Sqlite> =
|
||||||
|
QueryBuilder::new("insert into tasks(payload_key, payload, status_id)");
|
||||||
|
|
||||||
|
let args: crate::Result<Vec<(String, String)>> = tasks
|
||||||
|
.iter()
|
||||||
|
.map(|value| Ok((value.get_key(), serde_json::to_string(value.payload())?)))
|
||||||
|
.collect();
|
||||||
|
|
||||||
|
let mut affected_rows = 0;
|
||||||
|
// Chunk the query by the size limit of bind params
|
||||||
|
for chunk in args?.chunks(SQLITE_BIND_LIMIT / 3) {
|
||||||
|
builder.push_values(chunk, |mut builder, item| {
|
||||||
|
builder
|
||||||
|
.push_bind(&item.0)
|
||||||
|
.push_bind(&item.1)
|
||||||
|
.push_bind(TaskStatus::Pending);
|
||||||
|
});
|
||||||
|
builder.push("ON conflict (payload_key) DO NOTHING");
|
||||||
|
|
||||||
|
let query = builder.build();
|
||||||
|
|
||||||
|
affected_rows += query.execute(&mut *tx).await?.rows_affected();
|
||||||
|
builder.reset();
|
||||||
|
}
|
||||||
|
|
||||||
|
tx.commit().await?;
|
||||||
|
|
||||||
|
info!("{} rows inserted.", affected_rows);
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_tasks(&self, task_status: TaskStatus) -> impl Stream<Item = crate::Result<Task<T>>> {
|
||||||
|
let query= sqlx::query_as::<_, Task<T>>(
|
||||||
|
"
|
||||||
|
SELECT id, payload_key, payload, status_id, created_at, updated_at
|
||||||
|
FROM tasks
|
||||||
|
WHERE status_id = ?
|
||||||
|
ORDER BY created_at DESC
|
||||||
|
",
|
||||||
|
).bind(task_status);
|
||||||
|
|
||||||
|
query.fetch(&self.pool).err_into::<crate::Error>()
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn get_paginated_tasks(&self, page: &TaskPagination) -> crate::Result<TasksPage<T>> {
|
||||||
|
let mut builder: QueryBuilder<'_, sqlx::Sqlite> = QueryBuilder::new(
|
||||||
|
"select id, payload_key, payload, status_id, created_at, updated_at from tasks ",
|
||||||
|
);
|
||||||
|
|
||||||
|
if let Some(status) = &page.status {
|
||||||
|
builder.push("where status_id = ").push_bind(status);
|
||||||
|
}
|
||||||
|
|
||||||
|
builder.push("ORDER BY created_at DESC ");
|
||||||
|
|
||||||
|
if let Some(limit) = &page.offset {
|
||||||
|
builder.push("OFFSET ").push_bind(limit);
|
||||||
|
}
|
||||||
|
|
||||||
|
if let Some(limit) = &page.limit {
|
||||||
|
builder.push("LIMIT ").push_bind(limit);
|
||||||
|
}
|
||||||
|
|
||||||
|
let tasks = builder.build_query_as::<Task<T>>().fetch_all(&self.pool).await?;
|
||||||
|
|
||||||
|
Ok(TasksPage::new(tasks, page.clone()))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
@ -1,7 +1,7 @@
|
||||||
pub mod error;
|
pub mod error;
|
||||||
|
|
||||||
pub(crate) use error::*;
|
pub(crate) use error::*;
|
||||||
pub mod task_manager;
|
pub mod tasks;
|
||||||
mod database;
|
mod database;
|
||||||
|
|
||||||
pub fn add(left: u64, right: u64) -> u64 {
|
pub fn add(left: u64, right: u64) -> u64 {
|
||||||
|
|
|
||||||
|
|
@ -1,17 +1,11 @@
|
||||||
use crate::error::Error;
|
|
||||||
use chrono::Utc;
|
use chrono::Utc;
|
||||||
use directories::ProjectDirs;
|
use futures::StreamExt;
|
||||||
use futures::{StreamExt, TryStreamExt};
|
|
||||||
use serde::de::DeserializeOwned;
|
use serde::de::DeserializeOwned;
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
use sqlx::sqlite::{SqliteConnectOptions, SqliteJournalMode};
|
|
||||||
use sqlx::{QueryBuilder, Sqlite, SqlitePool};
|
|
||||||
use std::fmt::Display;
|
use std::fmt::Display;
|
||||||
use std::path::PathBuf;
|
|
||||||
use futures::stream::BoxStream;
|
|
||||||
use tabled::Tabled;
|
use tabled::Tabled;
|
||||||
use tokio::fs;
|
|
||||||
use tracing::{info, instrument};
|
mod manager;
|
||||||
|
|
||||||
#[derive(sqlx::Type, Debug, Clone)]
|
#[derive(sqlx::Type, Debug, Clone)]
|
||||||
#[repr(u8)]
|
#[repr(u8)]
|
||||||
|
|
@ -45,13 +39,19 @@ pub trait TaskPayloadKey {
|
||||||
fn get_key(&self) -> String;
|
fn get_key(&self) -> String;
|
||||||
}
|
}
|
||||||
|
|
||||||
pub trait TaskPayload: DeserializeOwned + Send + Unpin + 'static + Display + TaskPayloadKey {}
|
pub trait TaskPayload:
|
||||||
impl<T: DeserializeOwned + Send + Unpin + 'static + Display + TaskPayloadKey> TaskPayload for T {}
|
Serialize + DeserializeOwned + Send + Unpin + 'static + Display + TaskPayloadKey
|
||||||
|
{
|
||||||
|
}
|
||||||
|
impl<T: Serialize + DeserializeOwned + Send + Unpin + 'static + Display + TaskPayloadKey>
|
||||||
|
TaskPayload for T
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
pub type TaskJob<T: TaskPayload> = fn(&Task<T>) -> TaskStatus;
|
pub type TaskJob<T> = fn(&Task<T>) -> TaskStatus;
|
||||||
|
|
||||||
#[derive(sqlx::FromRow, Tabled, Debug)]
|
#[derive(sqlx::FromRow, Tabled, Debug)]
|
||||||
pub struct Task<T: DeserializeOwned + std::fmt::Display> {
|
pub struct Task<T: TaskPayload> {
|
||||||
id: u32,
|
id: u32,
|
||||||
payload_key: String,
|
payload_key: String,
|
||||||
#[sqlx(json)]
|
#[sqlx(json)]
|
||||||
|
|
@ -64,7 +64,13 @@ pub struct Task<T: DeserializeOwned + std::fmt::Display> {
|
||||||
updated_at: Option<chrono::DateTime<Utc>>,
|
updated_at: Option<chrono::DateTime<Utc>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<T: DeserializeOwned + std::fmt::Display> Task<T> {
|
impl<T: TaskPayload> Task<T> {
|
||||||
|
pub fn payload(&self) -> &T {
|
||||||
|
&self.payload
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T: TaskPayload> Task<T> {
|
||||||
pub fn get_key(&self) -> String {
|
pub fn get_key(&self) -> String {
|
||||||
self.payload_key.clone()
|
self.payload_key.clone()
|
||||||
}
|
}
|
||||||
|
|
@ -77,22 +83,3 @@ fn display_option_date(o: &Option<chrono::DateTime<Utc>>) -> String {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
struct TaskManager{}
|
|
||||||
|
|
||||||
impl TaskManager {
|
|
||||||
// pub async fn run_tasks<T: TaskPayload>(&self, func: TaskJob<T>) -> crate::Result<()> {
|
|
||||||
// let mut builder = Self::get_task_builder(Some(TaskStatus::Pending), None);
|
|
||||||
//
|
|
||||||
// let rows = builder.build_query_as::<Task<T>>().fetch(&self.pool);
|
|
||||||
//
|
|
||||||
// let result: Vec<(Task<T>, TaskStatus)> = rows.map(|x| {
|
|
||||||
// let task = x.unwrap();
|
|
||||||
// let status = func(&task);
|
|
||||||
//
|
|
||||||
// (task, status)
|
|
||||||
// }).collect().await;
|
|
||||||
//
|
|
||||||
// Ok(())
|
|
||||||
// }
|
|
||||||
}
|
|
||||||
32
lib_sync_core/src/tasks/manager.rs
Normal file
32
lib_sync_core/src/tasks/manager.rs
Normal file
|
|
@ -0,0 +1,32 @@
|
||||||
|
use futures::StreamExt;
|
||||||
|
use std::marker::PhantomData;
|
||||||
|
use crate::database::TaskStorage;
|
||||||
|
use crate::tasks::{Task, TaskJob, TaskPayload, TaskStatus};
|
||||||
|
|
||||||
|
struct TaskManager<S: TaskPayload, T: TaskStorage<S>>
|
||||||
|
{
|
||||||
|
storage: T,
|
||||||
|
_marker: PhantomData<S>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<S: TaskPayload, T: TaskStorage<S>> TaskManager<S, T> {
|
||||||
|
pub fn new(storage: T) -> Self {
|
||||||
|
Self {
|
||||||
|
storage,
|
||||||
|
_marker: PhantomData,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn run_tasks(&self, func: TaskJob<S>) -> crate::Result<()> {
|
||||||
|
let rows = self.storage.get_tasks(TaskStatus::Pending);
|
||||||
|
|
||||||
|
let result: Vec<(Task<S>, TaskStatus)> = rows.map(|x| {
|
||||||
|
let task = x.unwrap();
|
||||||
|
let status = func(&task);
|
||||||
|
|
||||||
|
(task, status)
|
||||||
|
}).collect().await;
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
}
|
||||||
Loading…
Add table
Add a link
Reference in a new issue