mirror of
https://github.com/meilisearch/meilisearch.git
synced 2024-11-25 19:45:05 +08:00
Do not store useless updates
This commit is contained in:
parent
ff5d3b59f5
commit
b2f4e67c9a
@ -137,9 +137,22 @@ impl<'p, 'pl: 'p> DocumentChanges<'p> for DocumentOperation<'pl> {
|
|||||||
(docid, vec![document_operation]),
|
(docid, vec![document_operation]),
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
// TODO clean the code to make sure we clean the useless operations
|
Some((_, offsets)) => {
|
||||||
// add a method to the MergeChanges trait
|
let useless_previous_addition = match self.index_documents_method {
|
||||||
Some((_, offsets)) => offsets.push(document_operation),
|
IndexDocumentsMethod::ReplaceDocuments => {
|
||||||
|
MergeDocumentForReplacement::USELESS_PREVIOUS_CHANGES
|
||||||
|
}
|
||||||
|
IndexDocumentsMethod::UpdateDocuments => {
|
||||||
|
MergeDocumentForUpdates::USELESS_PREVIOUS_CHANGES
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
if useless_previous_addition {
|
||||||
|
offsets.clear();
|
||||||
|
}
|
||||||
|
|
||||||
|
offsets.push(document_operation);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
previous_offset = iter.byte_offset();
|
previous_offset = iter.byte_offset();
|
||||||
@ -164,7 +177,10 @@ impl<'p, 'pl: 'p> DocumentChanges<'p> for DocumentOperation<'pl> {
|
|||||||
(docid, vec![InnerDocOp::Deletion]),
|
(docid, vec![InnerDocOp::Deletion]),
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
Some((_, offsets)) => offsets.push(InnerDocOp::Deletion),
|
Some((_, offsets)) => {
|
||||||
|
offsets.clear();
|
||||||
|
offsets.push(InnerDocOp::Deletion);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -176,10 +192,13 @@ impl<'p, 'pl: 'p> DocumentChanges<'p> for DocumentOperation<'pl> {
|
|||||||
// TODO We must drain the HashMap into a Vec because rayon::hash_map::IntoIter: !Clone
|
// TODO We must drain the HashMap into a Vec because rayon::hash_map::IntoIter: !Clone
|
||||||
let mut docids_version_offsets: Vec<_> = docids_version_offsets.drain().collect();
|
let mut docids_version_offsets: Vec<_> = docids_version_offsets.drain().collect();
|
||||||
// Reorder the offsets to make sure we iterate on the file sequentially
|
// Reorder the offsets to make sure we iterate on the file sequentially
|
||||||
match self.index_documents_method {
|
let sort_function_key = match self.index_documents_method {
|
||||||
Idm::ReplaceDocuments => MergeDocumentForReplacement::sort(&mut docids_version_offsets),
|
Idm::ReplaceDocuments => MergeDocumentForReplacement::sort_key,
|
||||||
Idm::UpdateDocuments => MergeDocumentForUpdates::sort(&mut docids_version_offsets),
|
Idm::UpdateDocuments => MergeDocumentForUpdates::sort_key,
|
||||||
}
|
};
|
||||||
|
|
||||||
|
// And finally sort them
|
||||||
|
docids_version_offsets.sort_unstable_by_key(|(_, (_, docops))| sort_function_key(docops));
|
||||||
|
|
||||||
Ok(docids_version_offsets
|
Ok(docids_version_offsets
|
||||||
.into_par_iter()
|
.into_par_iter()
|
||||||
@ -208,8 +227,11 @@ impl<'p, 'pl: 'p> DocumentChanges<'p> for DocumentOperation<'pl> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
trait MergeChanges {
|
trait MergeChanges {
|
||||||
/// Reorders the offsets to make sure we iterate on the file sequentially.
|
/// Wether the payloads in the list of operations are useless or not.
|
||||||
fn sort(changes_offsets: &mut [(CowStr, (DocumentId, Vec<InnerDocOp>))]);
|
const USELESS_PREVIOUS_CHANGES: bool;
|
||||||
|
|
||||||
|
/// Returns a key that is used to order the payloads the right way.
|
||||||
|
fn sort_key(docops: &[InnerDocOp]) -> usize;
|
||||||
|
|
||||||
fn merge(
|
fn merge(
|
||||||
rtxn: &RoTxn,
|
rtxn: &RoTxn,
|
||||||
@ -224,18 +246,15 @@ trait MergeChanges {
|
|||||||
struct MergeDocumentForReplacement;
|
struct MergeDocumentForReplacement;
|
||||||
|
|
||||||
impl MergeChanges for MergeDocumentForReplacement {
|
impl MergeChanges for MergeDocumentForReplacement {
|
||||||
|
const USELESS_PREVIOUS_CHANGES: bool = true;
|
||||||
|
|
||||||
/// Reorders to read only the last change.
|
/// Reorders to read only the last change.
|
||||||
fn sort(changes_offsets: &mut [(CowStr, (DocumentId, Vec<InnerDocOp>))]) {
|
fn sort_key(docops: &[InnerDocOp]) -> usize {
|
||||||
changes_offsets.sort_unstable_by_key(|(_, (_, offsets))| {
|
let f = |ido: &_| match ido {
|
||||||
offsets
|
|
||||||
.iter()
|
|
||||||
.rev()
|
|
||||||
.find_map(|ido| match ido {
|
|
||||||
InnerDocOp::Addition(add) => Some(add.content.as_ptr() as usize),
|
InnerDocOp::Addition(add) => Some(add.content.as_ptr() as usize),
|
||||||
InnerDocOp::Deletion => None,
|
InnerDocOp::Deletion => None,
|
||||||
})
|
};
|
||||||
.unwrap_or(0)
|
docops.iter().rev().find_map(f).unwrap_or(0)
|
||||||
});
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Returns only the most recent version of a document based on the updates from the payloads.
|
/// Returns only the most recent version of a document based on the updates from the payloads.
|
||||||
@ -295,17 +314,15 @@ impl MergeChanges for MergeDocumentForReplacement {
|
|||||||
struct MergeDocumentForUpdates;
|
struct MergeDocumentForUpdates;
|
||||||
|
|
||||||
impl MergeChanges for MergeDocumentForUpdates {
|
impl MergeChanges for MergeDocumentForUpdates {
|
||||||
|
const USELESS_PREVIOUS_CHANGES: bool = false;
|
||||||
|
|
||||||
/// Reorders to read the first changes first so that it's faster to read the first one and then the rest.
|
/// Reorders to read the first changes first so that it's faster to read the first one and then the rest.
|
||||||
fn sort(changes_offsets: &mut [(CowStr, (DocumentId, Vec<InnerDocOp>))]) {
|
fn sort_key(docops: &[InnerDocOp]) -> usize {
|
||||||
changes_offsets.sort_unstable_by_key(|(_, (_, offsets))| {
|
let f = |ido: &_| match ido {
|
||||||
offsets
|
|
||||||
.iter()
|
|
||||||
.find_map(|ido| match ido {
|
|
||||||
InnerDocOp::Addition(add) => Some(add.content.as_ptr() as usize),
|
InnerDocOp::Addition(add) => Some(add.content.as_ptr() as usize),
|
||||||
InnerDocOp::Deletion => None,
|
InnerDocOp::Deletion => None,
|
||||||
})
|
};
|
||||||
.unwrap_or(0)
|
docops.iter().find_map(f).unwrap_or(0)
|
||||||
});
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Reads the previous version of a document from the database, the new versions
|
/// Reads the previous version of a document from the database, the new versions
|
||||||
|
@ -1,4 +1,3 @@
|
|||||||
use std::fs::File;
|
|
||||||
use std::sync::RwLock;
|
use std::sync::RwLock;
|
||||||
use std::thread::{self, Builder};
|
use std::thread::{self, Builder};
|
||||||
|
|
||||||
@ -17,12 +16,10 @@ use super::document_change::DocumentChange;
|
|||||||
use super::extract::*;
|
use super::extract::*;
|
||||||
use super::merger::merge_grenad_entries;
|
use super::merger::merge_grenad_entries;
|
||||||
use super::StdResult;
|
use super::StdResult;
|
||||||
use crate::documents::{
|
use crate::documents::{PrimaryKey, DEFAULT_PRIMARY_KEY};
|
||||||
obkv_to_object, DocumentsBatchCursor, DocumentsBatchIndex, PrimaryKey, DEFAULT_PRIMARY_KEY,
|
|
||||||
};
|
|
||||||
use crate::update::new::channel::{DatabaseType, ExtractorSender};
|
use crate::update::new::channel::{DatabaseType, ExtractorSender};
|
||||||
use crate::update::GrenadParameters;
|
use crate::update::GrenadParameters;
|
||||||
use crate::{FieldsIdsMap, GlobalFieldsIdsMap, Index, InternalError, Result, UserError};
|
use crate::{FieldsIdsMap, GlobalFieldsIdsMap, Index, Result, UserError};
|
||||||
|
|
||||||
mod document_deletion;
|
mod document_deletion;
|
||||||
mod document_operation;
|
mod document_operation;
|
||||||
|
Loading…
Reference in New Issue
Block a user