Add enqueued and processed datetimes

This commit is contained in:
Clément Renault 2019-11-12 18:00:47 +01:00
parent a98949ff1d
commit acb5e624c6
16 changed files with 160 additions and 76 deletions

1
Cargo.lock generated
View File

@ -864,6 +864,7 @@ dependencies = [
"assert_matches 1.3.0 (registry+https://github.com/rust-lang/crates.io-index)", "assert_matches 1.3.0 (registry+https://github.com/rust-lang/crates.io-index)",
"bincode 1.2.0 (registry+https://github.com/rust-lang/crates.io-index)", "bincode 1.2.0 (registry+https://github.com/rust-lang/crates.io-index)",
"byteorder 1.3.2 (registry+https://github.com/rust-lang/crates.io-index)", "byteorder 1.3.2 (registry+https://github.com/rust-lang/crates.io-index)",
"chrono 0.4.9 (registry+https://github.com/rust-lang/crates.io-index)",
"crossbeam-channel 0.4.0 (registry+https://github.com/rust-lang/crates.io-index)", "crossbeam-channel 0.4.0 (registry+https://github.com/rust-lang/crates.io-index)",
"csv 1.1.1 (registry+https://github.com/rust-lang/crates.io-index)", "csv 1.1.1 (registry+https://github.com/rust-lang/crates.io-index)",
"deunicode 1.0.0 (registry+https://github.com/rust-lang/crates.io-index)", "deunicode 1.0.0 (registry+https://github.com/rust-lang/crates.io-index)",

View File

@ -8,6 +8,7 @@ edition = "2018"
arc-swap = "0.4.3" arc-swap = "0.4.3"
bincode = "1.1.4" bincode = "1.1.4"
byteorder = "1.3.2" byteorder = "1.3.2"
chrono = { version = "0.4.9", features = ["serde"] }
crossbeam-channel = "0.4.0" crossbeam-channel = "0.4.0"
deunicode = "1.0.0" deunicode = "1.0.0"
env_logger = "0.7.0" env_logger = "0.7.0"

View File

@ -323,7 +323,7 @@ mod tests {
let reader = env.read_txn().unwrap(); let reader = env.read_txn().unwrap();
let result = index.update_status(&reader, update_id).unwrap(); let result = index.update_status(&reader, update_id).unwrap();
assert_matches!(result, UpdateStatus::Processed { content } if content.error.is_none()); assert_matches!(result, Some(UpdateStatus::Processed { content }) if content.error.is_none());
} }
#[test] #[test]
@ -384,7 +384,7 @@ mod tests {
let reader = env.read_txn().unwrap(); let reader = env.read_txn().unwrap();
let result = index.update_status(&reader, update_id).unwrap(); let result = index.update_status(&reader, update_id).unwrap();
assert_matches!(result, UpdateStatus::Processed { content } if content.error.is_some()); assert_matches!(result, Some(UpdateStatus::Processed { content }) if content.error.is_some());
} }
#[test] #[test]
@ -434,7 +434,7 @@ mod tests {
let reader = env.read_txn().unwrap(); let reader = env.read_txn().unwrap();
let result = index.update_status(&reader, update_id).unwrap(); let result = index.update_status(&reader, update_id).unwrap();
assert_matches!(result, UpdateStatus::Processed { content } if content.error.is_none()); assert_matches!(result, Some(UpdateStatus::Processed { content }) if content.error.is_none());
} }
#[test] #[test]
@ -524,7 +524,7 @@ mod tests {
// check if it has been accepted // check if it has been accepted
let reader = env.read_txn().unwrap(); let reader = env.read_txn().unwrap();
let result = index.update_status(&reader, update_id).unwrap(); let result = index.update_status(&reader, update_id).unwrap();
assert_matches!(result, UpdateStatus::Processed { content } if content.error.is_none()); assert_matches!(result, Some(UpdateStatus::Processed { content }) if content.error.is_none());
reader.abort(); reader.abort();
let mut additions = index.documents_addition(); let mut additions = index.documents_addition();
@ -558,7 +558,7 @@ mod tests {
// check if it has been accepted // check if it has been accepted
let reader = env.read_txn().unwrap(); let reader = env.read_txn().unwrap();
let result = index.update_status(&reader, update_id).unwrap(); let result = index.update_status(&reader, update_id).unwrap();
assert_matches!(result, UpdateStatus::Processed { content } if content.error.is_none()); assert_matches!(result, Some(UpdateStatus::Processed { content }) if content.error.is_none());
// even try to search for a document // even try to search for a document
let results = index.query_builder().query(&reader, "21 ", 0..20).unwrap(); let results = index.query_builder().query(&reader, "21 ", 0..20).unwrap();
@ -604,7 +604,7 @@ mod tests {
// check if it has been accepted // check if it has been accepted
let reader = env.read_txn().unwrap(); let reader = env.read_txn().unwrap();
let result = index.update_status(&reader, update_id).unwrap(); let result = index.update_status(&reader, update_id).unwrap();
assert_matches!(result, UpdateStatus::Processed { content } if content.error.is_some()); assert_matches!(result, Some(UpdateStatus::Processed { content }) if content.error.is_some());
} }
#[test] #[test]
@ -668,7 +668,7 @@ mod tests {
let reader = env.read_txn().unwrap(); let reader = env.read_txn().unwrap();
let result = index.update_status(&reader, update_id).unwrap(); let result = index.update_status(&reader, update_id).unwrap();
assert_matches!(result, UpdateStatus::Processed { content } if content.error.is_none()); assert_matches!(result, Some(UpdateStatus::Processed { content }) if content.error.is_none());
let document: Option<IgnoredAny> = index.document(&reader, None, DocumentId(25)).unwrap(); let document: Option<IgnoredAny> = index.document(&reader, None, DocumentId(25)).unwrap();
assert!(document.is_none()); assert!(document.is_none());
@ -748,7 +748,7 @@ mod tests {
let reader = env.read_txn().unwrap(); let reader = env.read_txn().unwrap();
let result = index.update_status(&reader, update_id).unwrap(); let result = index.update_status(&reader, update_id).unwrap();
assert_matches!(result, UpdateStatus::Processed { content } if content.error.is_none()); assert_matches!(result, Some(UpdateStatus::Processed { content }) if content.error.is_none());
let document: Option<IgnoredAny> = index.document(&reader, None, DocumentId(25)).unwrap(); let document: Option<IgnoredAny> = index.document(&reader, None, DocumentId(25)).unwrap();
assert!(document.is_none()); assert!(document.is_none());
@ -791,7 +791,7 @@ mod tests {
let reader = env.read_txn().unwrap(); let reader = env.read_txn().unwrap();
let result = index.update_status(&reader, update_id).unwrap(); let result = index.update_status(&reader, update_id).unwrap();
assert_matches!(result, UpdateStatus::Processed { content } if content.error.is_none()); assert_matches!(result, Some(UpdateStatus::Processed { content }) if content.error.is_none());
let document: Option<serde_json::Value> = index let document: Option<serde_json::Value> = index
.document(&reader, None, DocumentId(7900334843754999545)) .document(&reader, None, DocumentId(7900334843754999545))

View File

@ -221,7 +221,7 @@ impl Index {
&self, &self,
reader: &heed::RoTxn, reader: &heed::RoTxn,
update_id: u64, update_id: u64,
) -> MResult<update::UpdateStatus> { ) -> MResult<Option<update::UpdateStatus>> {
update::update_status(reader, self.updates, self.updates_results, update_id) update::update_status(reader, self.updates, self.updates_results, update_id)
} }
@ -234,19 +234,21 @@ impl Index {
updates.reserve(last_id as usize); updates.reserve(last_id as usize);
for id in 0..=last_id { for id in 0..=last_id {
let update = self.update_status(reader, id)?; if let Some(update) = self.update_status(reader, id)? {
updates.push(update); updates.push(update);
last_update_result_id = id; last_update_result_id = id;
} }
} }
}
// retrieve all enqueued updates // retrieve all enqueued updates
if let Some((last_id, _)) = self.updates.last_update_id(reader)? { if let Some((last_id, _)) = self.updates.last_update_id(reader)? {
for id in last_update_result_id + 1..last_id { for id in last_update_result_id + 1..last_id {
let update = self.update_status(reader, id)?; if let Some(update) = self.update_status(reader, id)? {
updates.push(update); updates.push(update);
} }
} }
}
Ok(updates) Ok(updates)
} }

View File

@ -5,8 +5,7 @@ use heed::Result as ZResult;
#[derive(Copy, Clone)] #[derive(Copy, Clone)]
pub struct UpdatesResults { pub struct UpdatesResults {
pub(crate) updates_results: pub(crate) updates_results: heed::Database<OwnedType<BEU64>, SerdeJson<ProcessedUpdateResult>>,
heed::Database<OwnedType<BEU64>, SerdeJson<ProcessedUpdateResult>>,
} }
impl UpdatesResults { impl UpdatesResults {

View File

@ -26,7 +26,7 @@ pub fn push_clear_all(
updates_results_store: store::UpdatesResults, updates_results_store: store::UpdatesResults,
) -> MResult<u64> { ) -> MResult<u64> {
let last_update_id = next_update_id(writer, updates_store, updates_results_store)?; let last_update_id = next_update_id(writer, updates_store, updates_results_store)?;
let update = Update::ClearAll; let update = Update::clear_all();
updates_store.put_update(writer, last_update_id, &update)?; updates_store.put_update(writer, last_update_id, &update)?;
Ok(last_update_id) Ok(last_update_id)

View File

@ -18,7 +18,7 @@ pub fn push_customs_update(
) -> ZResult<u64> { ) -> ZResult<u64> {
let last_update_id = next_update_id(writer, updates_store, updates_results_store)?; let last_update_id = next_update_id(writer, updates_store, updates_results_store)?;
let update = Update::Customs(customs); let update = Update::customs(customs);
updates_store.put_update(writer, last_update_id, &update)?; updates_store.put_update(writer, last_update_id, &update)?;
Ok(last_update_id) Ok(last_update_id)

View File

@ -91,9 +91,9 @@ pub fn push_documents_addition<D: serde::Serialize>(
let last_update_id = next_update_id(writer, updates_store, updates_results_store)?; let last_update_id = next_update_id(writer, updates_store, updates_results_store)?;
let update = if is_partial { let update = if is_partial {
Update::DocumentsPartial(values) Update::documents_partial(values)
} else { } else {
Update::DocumentsAddition(values) Update::documents_addition(values)
}; };
updates_store.put_update(writer, last_update_id, &update)?; updates_store.put_update(writer, last_update_id, &update)?;

View File

@ -76,7 +76,7 @@ pub fn push_documents_deletion(
) -> MResult<u64> { ) -> MResult<u64> {
let last_update_id = next_update_id(writer, updates_store, updates_results_store)?; let last_update_id = next_update_id(writer, updates_store, updates_results_store)?;
let update = Update::DocumentsDeletion(deletion); let update = Update::documents_deletion(deletion);
updates_store.put_update(writer, last_update_id, &update)?; updates_store.put_update(writer, last_update_id, &update)?;
Ok(last_update_id) Ok(last_update_id)

View File

@ -24,6 +24,7 @@ use std::cmp;
use std::collections::{BTreeMap, BTreeSet, HashMap}; use std::collections::{BTreeMap, BTreeSet, HashMap};
use std::time::Instant; use std::time::Instant;
use chrono::{DateTime, Utc};
use heed::Result as ZResult; use heed::Result as ZResult;
use log::debug; use log::debug;
use serde::{Deserialize, Serialize}; use serde::{Deserialize, Serialize};
@ -32,7 +33,85 @@ use crate::{store, DocumentId, MResult};
use meilidb_schema::Schema; use meilidb_schema::Schema;
#[derive(Debug, Clone, Serialize, Deserialize)] #[derive(Debug, Clone, Serialize, Deserialize)]
pub enum Update { pub struct Update {
data: UpdateData,
enqueued_at: DateTime<Utc>,
}
impl Update {
fn clear_all() -> Update {
Update {
data: UpdateData::ClearAll,
enqueued_at: Utc::now(),
}
}
fn schema(data: Schema) -> Update {
Update {
data: UpdateData::Schema(data),
enqueued_at: Utc::now(),
}
}
fn customs(data: Vec<u8>) -> Update {
Update {
data: UpdateData::Customs(data),
enqueued_at: Utc::now(),
}
}
fn documents_addition(data: Vec<HashMap<String, serde_json::Value>>) -> Update {
Update {
data: UpdateData::DocumentsAddition(data),
enqueued_at: Utc::now(),
}
}
fn documents_partial(data: Vec<HashMap<String, serde_json::Value>>) -> Update {
Update {
data: UpdateData::DocumentsPartial(data),
enqueued_at: Utc::now(),
}
}
fn documents_deletion(data: Vec<DocumentId>) -> Update {
Update {
data: UpdateData::DocumentsDeletion(data),
enqueued_at: Utc::now(),
}
}
fn synonyms_addition(data: BTreeMap<String, Vec<String>>) -> Update {
Update {
data: UpdateData::SynonymsAddition(data),
enqueued_at: Utc::now(),
}
}
fn synonyms_deletion(data: BTreeMap<String, Option<Vec<String>>>) -> Update {
Update {
data: UpdateData::SynonymsDeletion(data),
enqueued_at: Utc::now(),
}
}
fn stop_words_addition(data: BTreeSet<String>) -> Update {
Update {
data: UpdateData::StopWordsAddition(data),
enqueued_at: Utc::now(),
}
}
fn stop_words_deletion(data: BTreeSet<String>) -> Update {
Update {
data: UpdateData::StopWordsDeletion(data),
enqueued_at: Utc::now(),
}
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
pub enum UpdateData {
ClearAll, ClearAll,
Schema(Schema), Schema(Schema),
Customs(Vec<u8>), Customs(Vec<u8>),
@ -45,31 +124,31 @@ pub enum Update {
StopWordsDeletion(BTreeSet<String>), StopWordsDeletion(BTreeSet<String>),
} }
impl Update { impl UpdateData {
pub fn update_type(&self) -> UpdateType { pub fn update_type(&self) -> UpdateType {
match self { match self {
Update::ClearAll => UpdateType::ClearAll, UpdateData::ClearAll => UpdateType::ClearAll,
Update::Schema(_) => UpdateType::Schema, UpdateData::Schema(_) => UpdateType::Schema,
Update::Customs(_) => UpdateType::Customs, UpdateData::Customs(_) => UpdateType::Customs,
Update::DocumentsAddition(addition) => UpdateType::DocumentsAddition { UpdateData::DocumentsAddition(addition) => UpdateType::DocumentsAddition {
number: addition.len(), number: addition.len(),
}, },
Update::DocumentsPartial(addition) => UpdateType::DocumentsPartial { UpdateData::DocumentsPartial(addition) => UpdateType::DocumentsPartial {
number: addition.len(), number: addition.len(),
}, },
Update::DocumentsDeletion(deletion) => UpdateType::DocumentsDeletion { UpdateData::DocumentsDeletion(deletion) => UpdateType::DocumentsDeletion {
number: deletion.len(), number: deletion.len(),
}, },
Update::SynonymsAddition(addition) => UpdateType::SynonymsAddition { UpdateData::SynonymsAddition(addition) => UpdateType::SynonymsAddition {
number: addition.len(), number: addition.len(),
}, },
Update::SynonymsDeletion(deletion) => UpdateType::SynonymsDeletion { UpdateData::SynonymsDeletion(deletion) => UpdateType::SynonymsDeletion {
number: deletion.len(), number: deletion.len(),
}, },
Update::StopWordsAddition(addition) => UpdateType::StopWordsAddition { UpdateData::StopWordsAddition(addition) => UpdateType::StopWordsAddition {
number: addition.len(), number: addition.len(),
}, },
Update::StopWordsDeletion(deletion) => UpdateType::StopWordsDeletion { UpdateData::StopWordsDeletion(deletion) => UpdateType::StopWordsDeletion {
number: deletion.len(), number: deletion.len(),
}, },
} }
@ -99,26 +178,28 @@ pub struct ProcessedUpdateResult {
#[serde(skip_serializing_if = "Option::is_none")] #[serde(skip_serializing_if = "Option::is_none")]
pub error: Option<String>, pub error: Option<String>,
pub duration: f64, // in seconds pub duration: f64, // in seconds
pub enqueued_at: DateTime<Utc>,
pub processed_at: DateTime<Utc>,
} }
#[derive(Debug, Clone, Serialize, Deserialize)] #[derive(Debug, Clone, Serialize, Deserialize)]
pub struct EnqueuedUpdateResult { pub struct EnqueuedUpdateResult {
pub update_id: u64, pub update_id: u64,
pub update_type: UpdateType, pub update_type: UpdateType,
pub enqueued_at: DateTime<Utc>,
} }
#[derive(Debug, Clone, Serialize, Deserialize)] #[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(tag = "status")] #[serde(rename_all = "camelCase", tag = "status")]
pub enum UpdateStatus { pub enum UpdateStatus {
Enqueued { Enqueued {
#[serde(flatten)] #[serde(flatten)]
content: EnqueuedUpdateResult content: EnqueuedUpdateResult,
}, },
Processed { Processed {
#[serde(flatten)] #[serde(flatten)]
content: ProcessedUpdateResult content: ProcessedUpdateResult,
}, },
Unknown,
} }
pub fn update_status( pub fn update_status(
@ -126,19 +207,19 @@ pub fn update_status(
updates_store: store::Updates, updates_store: store::Updates,
updates_results_store: store::UpdatesResults, updates_results_store: store::UpdatesResults,
update_id: u64, update_id: u64,
) -> MResult<UpdateStatus> { ) -> MResult<Option<UpdateStatus>> {
match updates_results_store.update_result(reader, update_id)? { match updates_results_store.update_result(reader, update_id)? {
Some(result) => Ok(UpdateStatus::Processed { content: result }), Some(result) => Ok(Some(UpdateStatus::Processed { content: result })),
None => { None => match updates_store.get(reader, update_id)? {
if let Some(update) = updates_store.get(reader, update_id)? { Some(update) => Ok(Some(UpdateStatus::Enqueued {
Ok(UpdateStatus::Enqueued { content: EnqueuedUpdateResult { content: EnqueuedUpdateResult {
update_id, update_id,
update_type: update.update_type(), update_type: update.data.update_type(),
}}) enqueued_at: update.enqueued_at,
} else { },
Ok(UpdateStatus::Unknown) })),
} None => Ok(None),
} },
} }
} }
@ -167,8 +248,10 @@ pub fn update_task<'a, 'b>(
) -> MResult<ProcessedUpdateResult> { ) -> MResult<ProcessedUpdateResult> {
debug!("Processing update number {}", update_id); debug!("Processing update number {}", update_id);
let (update_type, result, duration) = match update { let Update { enqueued_at, data } = update;
Update::ClearAll => {
let (update_type, result, duration) = match data {
UpdateData::ClearAll => {
let start = Instant::now(); let start = Instant::now();
let update_type = UpdateType::ClearAll; let update_type = UpdateType::ClearAll;
@ -183,7 +266,7 @@ pub fn update_task<'a, 'b>(
(update_type, result, start.elapsed()) (update_type, result, start.elapsed())
} }
Update::Schema(schema) => { UpdateData::Schema(schema) => {
let start = Instant::now(); let start = Instant::now();
let update_type = UpdateType::Schema; let update_type = UpdateType::Schema;
@ -199,7 +282,7 @@ pub fn update_task<'a, 'b>(
(update_type, result, start.elapsed()) (update_type, result, start.elapsed())
} }
Update::Customs(customs) => { UpdateData::Customs(customs) => {
let start = Instant::now(); let start = Instant::now();
let update_type = UpdateType::Customs; let update_type = UpdateType::Customs;
@ -207,7 +290,7 @@ pub fn update_task<'a, 'b>(
(update_type, result, start.elapsed()) (update_type, result, start.elapsed())
} }
Update::DocumentsAddition(documents) => { UpdateData::DocumentsAddition(documents) => {
let start = Instant::now(); let start = Instant::now();
let update_type = UpdateType::DocumentsAddition { let update_type = UpdateType::DocumentsAddition {
@ -226,7 +309,7 @@ pub fn update_task<'a, 'b>(
(update_type, result, start.elapsed()) (update_type, result, start.elapsed())
} }
Update::DocumentsPartial(documents) => { UpdateData::DocumentsPartial(documents) => {
let start = Instant::now(); let start = Instant::now();
let update_type = UpdateType::DocumentsPartial { let update_type = UpdateType::DocumentsPartial {
@ -245,7 +328,7 @@ pub fn update_task<'a, 'b>(
(update_type, result, start.elapsed()) (update_type, result, start.elapsed())
} }
Update::DocumentsDeletion(documents) => { UpdateData::DocumentsDeletion(documents) => {
let start = Instant::now(); let start = Instant::now();
let update_type = UpdateType::DocumentsDeletion { let update_type = UpdateType::DocumentsDeletion {
@ -264,7 +347,7 @@ pub fn update_task<'a, 'b>(
(update_type, result, start.elapsed()) (update_type, result, start.elapsed())
} }
Update::SynonymsAddition(synonyms) => { UpdateData::SynonymsAddition(synonyms) => {
let start = Instant::now(); let start = Instant::now();
let update_type = UpdateType::SynonymsAddition { let update_type = UpdateType::SynonymsAddition {
@ -275,7 +358,7 @@ pub fn update_task<'a, 'b>(
(update_type, result, start.elapsed()) (update_type, result, start.elapsed())
} }
Update::SynonymsDeletion(synonyms) => { UpdateData::SynonymsDeletion(synonyms) => {
let start = Instant::now(); let start = Instant::now();
let update_type = UpdateType::SynonymsDeletion { let update_type = UpdateType::SynonymsDeletion {
@ -286,7 +369,7 @@ pub fn update_task<'a, 'b>(
(update_type, result, start.elapsed()) (update_type, result, start.elapsed())
} }
Update::StopWordsAddition(stop_words) => { UpdateData::StopWordsAddition(stop_words) => {
let start = Instant::now(); let start = Instant::now();
let update_type = UpdateType::StopWordsAddition { let update_type = UpdateType::StopWordsAddition {
@ -298,7 +381,7 @@ pub fn update_task<'a, 'b>(
(update_type, result, start.elapsed()) (update_type, result, start.elapsed())
} }
Update::StopWordsDeletion(stop_words) => { UpdateData::StopWordsDeletion(stop_words) => {
let start = Instant::now(); let start = Instant::now();
let update_type = UpdateType::StopWordsDeletion { let update_type = UpdateType::StopWordsDeletion {
@ -329,6 +412,8 @@ pub fn update_task<'a, 'b>(
update_type, update_type,
error: result.map_err(|e| e.to_string()).err(), error: result.map_err(|e| e.to_string()).err(),
duration: duration.as_secs_f64(), duration: duration.as_secs_f64(),
enqueued_at,
processed_at: Utc::now(),
}; };
Ok(status) Ok(status)

View File

@ -68,7 +68,7 @@ pub fn push_schema_update(
) -> MResult<u64> { ) -> MResult<u64> {
let last_update_id = next_update_id(writer, updates_store, updates_results_store)?; let last_update_id = next_update_id(writer, updates_store, updates_results_store)?;
let update = Update::Schema(schema); let update = Update::schema(schema);
updates_store.put_update(writer, last_update_id, &update)?; updates_store.put_update(writer, last_update_id, &update)?;
Ok(last_update_id) Ok(last_update_id)

View File

@ -53,7 +53,7 @@ pub fn push_stop_words_addition(
) -> MResult<u64> { ) -> MResult<u64> {
let last_update_id = next_update_id(writer, updates_store, updates_results_store)?; let last_update_id = next_update_id(writer, updates_store, updates_results_store)?;
let update = Update::StopWordsAddition(addition); let update = Update::stop_words_addition(addition);
updates_store.put_update(writer, last_update_id, &update)?; updates_store.put_update(writer, last_update_id, &update)?;
Ok(last_update_id) Ok(last_update_id)

View File

@ -54,7 +54,7 @@ pub fn push_stop_words_deletion(
) -> MResult<u64> { ) -> MResult<u64> {
let last_update_id = next_update_id(writer, updates_store, updates_results_store)?; let last_update_id = next_update_id(writer, updates_store, updates_results_store)?;
let update = Update::StopWordsDeletion(deletion); let update = Update::stop_words_deletion(deletion);
updates_store.put_update(writer, last_update_id, &update)?; updates_store.put_update(writer, last_update_id, &update)?;
Ok(last_update_id) Ok(last_update_id)

View File

@ -63,7 +63,7 @@ pub fn push_synonyms_addition(
) -> MResult<u64> { ) -> MResult<u64> {
let last_update_id = next_update_id(writer, updates_store, updates_results_store)?; let last_update_id = next_update_id(writer, updates_store, updates_results_store)?;
let update = Update::SynonymsAddition(addition); let update = Update::synonyms_addition(addition);
updates_store.put_update(writer, last_update_id, &update)?; updates_store.put_update(writer, last_update_id, &update)?;
Ok(last_update_id) Ok(last_update_id)

View File

@ -70,7 +70,7 @@ pub fn push_synonyms_deletion(
) -> MResult<u64> { ) -> MResult<u64> {
let last_update_id = next_update_id(writer, updates_store, updates_results_store)?; let last_update_id = next_update_id(writer, updates_store, updates_results_store)?;
let update = Update::SynonymsDeletion(deletion); let update = Update::synonyms_deletion(deletion);
updates_store.put_update(writer, last_update_id, &update)?; updates_store.put_update(writer, last_update_id, &update)?;
Ok(last_update_id) Ok(last_update_id)

View File

@ -1,5 +1,5 @@
use http::StatusCode; use http::StatusCode;
use meilidb_core::{ProcessedUpdateResult, UpdateStatus}; use meilidb_core::ProcessedUpdateResult;
use meilidb_schema::Schema; use meilidb_schema::Schema;
use serde_json::json; use serde_json::json;
use tide::response::IntoResponse; use tide::response::IntoResponse;
@ -150,16 +150,12 @@ pub async fn get_update_status(ctx: Context<Data>) -> SResult<Response> {
.map_err(ResponseError::internal)?; .map_err(ResponseError::internal)?;
let response = match status { let response = match status {
UpdateStatus::Unknown => { Some(status) => tide::response::json(status)
tide::response::json(json!({ "message": "unknown update id" }))
.with_status(StatusCode::NOT_FOUND)
.into_response()
}
status => {
tide::response::json(status)
.with_status(StatusCode::OK) .with_status(StatusCode::OK)
.into_response() .into_response(),
} None => tide::response::json(json!({ "message": "unknown update id" }))
.with_status(StatusCode::NOT_FOUND)
.into_response(),
}; };
Ok(response) Ok(response)