mirror of
https://github.com/meilisearch/meilisearch.git
synced 2025-01-30 23:13:09 +08:00
Implement the document merge function for the update method
This commit is contained in:
parent
c683fa98e6
commit
637a9c8bdd
@ -72,6 +72,19 @@ impl<R> DocumentsBatchCursor<R> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl<R: io::Read + io::Seek> DocumentsBatchCursor<R> {
|
impl<R: io::Read + io::Seek> DocumentsBatchCursor<R> {
|
||||||
|
/// Returns a single document from the database.
|
||||||
|
pub fn get(
|
||||||
|
&mut self,
|
||||||
|
offset: u32,
|
||||||
|
) -> Result<Option<KvReader<'_, FieldId>>, DocumentsBatchCursorError> {
|
||||||
|
match self.cursor.move_on_key_equal_to(offset.to_be_bytes())? {
|
||||||
|
Some((key, value)) if key != DOCUMENTS_BATCH_INDEX_KEY => {
|
||||||
|
Ok(Some(KvReader::new(value)))
|
||||||
|
}
|
||||||
|
_otherwise => Ok(None),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// Returns the next document, starting from the first one. Subsequent calls to
|
/// Returns the next document, starting from the first one. Subsequent calls to
|
||||||
/// `next_document` advance the document reader until all the documents have been read.
|
/// `next_document` advance the document reader until all the documents have been read.
|
||||||
pub fn next_document(
|
pub fn next_document(
|
||||||
|
@ -40,7 +40,11 @@ impl DocumentChange {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl Deletion {
|
impl Deletion {
|
||||||
pub fn new(docid: DocumentId, external_docid: String, current: Box<KvReaderFieldId>) -> Self {
|
pub fn create(
|
||||||
|
docid: DocumentId,
|
||||||
|
external_docid: String,
|
||||||
|
current: Box<KvReaderFieldId>,
|
||||||
|
) -> Self {
|
||||||
Self { docid, external_docid, current }
|
Self { docid, external_docid, current }
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -54,6 +58,10 @@ impl Deletion {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl Insertion {
|
impl Insertion {
|
||||||
|
pub fn create(docid: DocumentId, external_docid: String, new: Box<KvReaderFieldId>) -> Self {
|
||||||
|
Insertion { docid, external_docid, new }
|
||||||
|
}
|
||||||
|
|
||||||
fn docid(&self) -> DocumentId {
|
fn docid(&self) -> DocumentId {
|
||||||
self.docid
|
self.docid
|
||||||
}
|
}
|
||||||
@ -64,6 +72,15 @@ impl Insertion {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl Update {
|
impl Update {
|
||||||
|
pub fn create(
|
||||||
|
docid: DocumentId,
|
||||||
|
external_docid: String,
|
||||||
|
current: Box<KvReaderFieldId>,
|
||||||
|
new: Box<KvReaderFieldId>,
|
||||||
|
) -> Self {
|
||||||
|
Update { docid, external_docid, current, new }
|
||||||
|
}
|
||||||
|
|
||||||
fn docid(&self) -> DocumentId {
|
fn docid(&self) -> DocumentId {
|
||||||
self.docid
|
self.docid
|
||||||
}
|
}
|
||||||
|
@ -5,19 +5,21 @@ mod items_pool;
|
|||||||
mod global_fields_ids_map;
|
mod global_fields_ids_map;
|
||||||
|
|
||||||
mod indexer {
|
mod indexer {
|
||||||
|
use std::borrow::Cow;
|
||||||
use std::collections::{BTreeMap, HashMap};
|
use std::collections::{BTreeMap, HashMap};
|
||||||
use std::fs::File;
|
use std::fs::File;
|
||||||
use std::io::Cursor;
|
use std::io::Cursor;
|
||||||
use std::os::unix::fs::MetadataExt;
|
use std::os::unix::fs::MetadataExt;
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use heed::types::Bytes;
|
||||||
use heed::RoTxn;
|
use heed::RoTxn;
|
||||||
use memmap2::Mmap;
|
use memmap2::Mmap;
|
||||||
use rayon::iter::{IntoParallelIterator, ParallelBridge, ParallelIterator};
|
use rayon::iter::{IntoParallelIterator, ParallelBridge, ParallelIterator};
|
||||||
use roaring::RoaringBitmap;
|
use roaring::RoaringBitmap;
|
||||||
use serde_json::Value;
|
use serde_json::Value;
|
||||||
|
|
||||||
use super::document_change::{self, DocumentChange};
|
use super::document_change::{self, DocumentChange, Insertion, Update};
|
||||||
use super::items_pool::ItemsPool;
|
use super::items_pool::ItemsPool;
|
||||||
use crate::documents::{
|
use crate::documents::{
|
||||||
obkv_to_object, DocumentIdExtractionError, DocumentsBatchReader, PrimaryKey,
|
obkv_to_object, DocumentIdExtractionError, DocumentsBatchReader, PrimaryKey,
|
||||||
@ -28,7 +30,7 @@ mod indexer {
|
|||||||
};
|
};
|
||||||
|
|
||||||
pub type KvReaderFieldId = obkv2::KvReader<FieldId>;
|
pub type KvReaderFieldId = obkv2::KvReader<FieldId>;
|
||||||
pub type KvWriterFieldId<W> = obkv2::KvWriter<FieldId, W>;
|
pub type KvWriterFieldId<W> = obkv2::KvWriter<W, FieldId>;
|
||||||
|
|
||||||
pub struct DocumentOperationIndexer {
|
pub struct DocumentOperationIndexer {
|
||||||
operations: Vec<Payload>,
|
operations: Vec<Payload>,
|
||||||
@ -293,9 +295,13 @@ mod indexer {
|
|||||||
)
|
)
|
||||||
.into()),
|
.into()),
|
||||||
}?;
|
}?;
|
||||||
Ok(DocumentChange::Deletion(document_change::Deletion::new(
|
|
||||||
|
/// TODO create a function for this
|
||||||
|
let document = document.as_bytes().to_vec().into_boxed_slice().into();
|
||||||
|
Ok(DocumentChange::Deletion(document_change::Deletion::create(
|
||||||
docid,
|
docid,
|
||||||
external_docid,
|
external_docid,
|
||||||
|
document,
|
||||||
)))
|
)))
|
||||||
})
|
})
|
||||||
}))
|
}))
|
||||||
@ -358,14 +364,13 @@ mod indexer {
|
|||||||
external_docid: String,
|
external_docid: String,
|
||||||
operations: &[DocumentOperation],
|
operations: &[DocumentOperation],
|
||||||
) -> Result<Option<DocumentChange>> {
|
) -> Result<Option<DocumentChange>> {
|
||||||
let mut document = BTreeMap::new();
|
let mut document = BTreeMap::<_, Cow<_>>::new();
|
||||||
let original_obkv =
|
let original = index.documents.remap_data_type::<Bytes>().get(rtxn, &docid)?;
|
||||||
index.documents.remap_data_type::<heed::types::Bytes>().get(rtxn, &docid)?;
|
let original: Option<&KvReaderFieldId> = original.map(Into::into);
|
||||||
let original_obkv: Option<&KvReaderFieldId> = original_obkv.map(Into::into);
|
|
||||||
|
|
||||||
if let Some(original_obkv) = original_obkv {
|
if let Some(original) = original {
|
||||||
original_obkv.into_iter().for_each(|(k, v)| {
|
original.into_iter().for_each(|(k, v)| {
|
||||||
document.insert(k, v.to_vec());
|
document.insert(k, v.into());
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -376,10 +381,10 @@ mod indexer {
|
|||||||
let operations = &operations[last_deletion.map_or(0, |i| i + 1)..];
|
let operations = &operations[last_deletion.map_or(0, |i| i + 1)..];
|
||||||
|
|
||||||
if operations.is_empty() {
|
if operations.is_empty() {
|
||||||
match original_obkv {
|
match original {
|
||||||
Some(original_obkv) => {
|
Some(original_obkv) => {
|
||||||
let current = original_obkv.as_bytes().to_vec().into_boxed_slice().into();
|
let current = original_obkv.as_bytes().to_vec().into_boxed_slice().into();
|
||||||
return Ok(Some(DocumentChange::Deletion(document_change::Deletion::new(
|
return Ok(Some(DocumentChange::Deletion(document_change::Deletion::create(
|
||||||
docid,
|
docid,
|
||||||
external_docid,
|
external_docid,
|
||||||
current,
|
current,
|
||||||
@ -390,25 +395,38 @@ mod indexer {
|
|||||||
}
|
}
|
||||||
|
|
||||||
for operation in operations {
|
for operation in operations {
|
||||||
let DocumentOffset { content, offset } = ;
|
let DocumentOffset { content, offset } = match operation {
|
||||||
|
DocumentOperation::Addition(offset) => offset,
|
||||||
|
DocumentOperation::Deletion => unreachable!("Deletion in document operations"),
|
||||||
|
};
|
||||||
|
|
||||||
let reader = DocumentsBatchReader::from_reader(Cursor::new(content.as_ref()))?;
|
let reader = DocumentsBatchReader::from_reader(Cursor::new(content.as_ref()))?;
|
||||||
let (mut cursor, batch_index) = reader.into_cursor_and_fields_index();
|
let (mut cursor, batch_index) = reader.into_cursor_and_fields_index();
|
||||||
let obkv = cursor.get(*offset)?.expect("must exists");
|
let update = cursor.get(*offset)?.expect("must exists");
|
||||||
|
|
||||||
obkv.into_iter().for_each(|(k, v)| {
|
update.into_iter().for_each(|(k, v)| {
|
||||||
let field_name = batch_index.name(k).unwrap();
|
let field_name = batch_index.name(k).unwrap();
|
||||||
let id = fields_ids_map.id(field_name).unwrap();
|
let id = fields_ids_map.id(field_name).unwrap();
|
||||||
document.insert(id, v.to_vec());
|
document.insert(id, v.to_vec().into());
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut writer = KvWriterFieldId::memory();
|
let mut writer = KvWriterFieldId::memory();
|
||||||
document.into_iter().for_each(|(id, value)| writer.insert(id, value).unwrap());
|
document.into_iter().for_each(|(id, value)| writer.insert(id, value).unwrap());
|
||||||
let boxed = writer.into_inner().unwrap().into_boxed_slice();
|
/// TODO create a function for this conversion
|
||||||
|
let new = writer.into_inner().unwrap().into_boxed_slice().into();
|
||||||
|
|
||||||
// Box<KvReaderFieldId>
|
match original {
|
||||||
|
Some(original) => {
|
||||||
Ok(boxed.into())
|
/// TODO create a function for this conversion
|
||||||
|
let current = original.as_bytes().to_vec().into_boxed_slice().into();
|
||||||
|
let update = Update::create(docid, external_docid, current, new);
|
||||||
|
Ok(Some(DocumentChange::Update(update)))
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
let insertion = Insertion::create(docid, external_docid, new);
|
||||||
|
Ok(Some(DocumentChange::Insertion(insertion)))
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
Loading…
x
Reference in New Issue
Block a user