mirror of
https://github.com/meilisearch/meilisearch.git
synced 2024-11-23 10:37:41 +08:00
Fix multiple bugs
This commit is contained in:
parent
4e86ecf807
commit
0a83a86e65
2
Cargo.lock
generated
2
Cargo.lock
generated
@ -1022,7 +1022,7 @@ checksum = "2839e79665f131bdb5782e51f2c6c9599c133c6098982a54c794358bf432529c"
|
|||||||
[[package]]
|
[[package]]
|
||||||
name = "oxidized-mtbl"
|
name = "oxidized-mtbl"
|
||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
source = "git+https://github.com/Kerollmops/oxidized-mtbl.git?rev=6acef3d#6acef3d0fc7fec6a3701038860e51f8bbcee1ee6"
|
source = "git+https://github.com/Kerollmops/oxidized-mtbl.git?rev=9451be8#9451be8829562f7d1f8d34aa3ecb81c5106a0623"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"byteorder 1.3.4",
|
"byteorder 1.3.4",
|
||||||
"crc32c",
|
"crc32c",
|
||||||
|
@ -18,7 +18,7 @@ jemallocator = "0.3.2"
|
|||||||
levenshtein_automata = { version = "0.2.0", features = ["fst_automaton"] }
|
levenshtein_automata = { version = "0.2.0", features = ["fst_automaton"] }
|
||||||
memmap = "0.7.0"
|
memmap = "0.7.0"
|
||||||
once_cell = "1.4.0"
|
once_cell = "1.4.0"
|
||||||
oxidized-mtbl = { git = "https://github.com/Kerollmops/oxidized-mtbl.git", rev = "6acef3d" }
|
oxidized-mtbl = { git = "https://github.com/Kerollmops/oxidized-mtbl.git", rev = "9451be8" }
|
||||||
rayon = "1.3.0"
|
rayon = "1.3.0"
|
||||||
roaring = "0.5.2"
|
roaring = "0.5.2"
|
||||||
slice-group-by = "0.2.6"
|
slice-group-by = "0.2.6"
|
||||||
|
@ -3,6 +3,9 @@ use std::time::Instant;
|
|||||||
|
|
||||||
use pathfinding::directed::dijkstra::dijkstra;
|
use pathfinding::directed::dijkstra::dijkstra;
|
||||||
|
|
||||||
|
use smallvec::smallvec; // the macro
|
||||||
|
use crate::SmallVec16;
|
||||||
|
|
||||||
const ONE_ATTRIBUTE: u32 = 1000;
|
const ONE_ATTRIBUTE: u32 = 1000;
|
||||||
const MAX_DISTANCE: u32 = 8;
|
const MAX_DISTANCE: u32 = 8;
|
||||||
|
|
||||||
@ -27,17 +30,17 @@ fn extract_position(position: u32) -> (u32, u32) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone, PartialOrd, Ord, PartialEq, Eq, Hash)]
|
#[derive(Debug, Clone, PartialOrd, Ord, PartialEq, Eq, Hash)]
|
||||||
struct Path(Vec<u32>);
|
struct Path(SmallVec16<u32>);
|
||||||
|
|
||||||
impl Path {
|
impl Path {
|
||||||
fn new(positions: &[Vec<u32>]) -> Option<Path> {
|
fn new(positions: &[Vec<u32>]) -> Option<Path> {
|
||||||
let position = positions.first()?.first()?;
|
let position = positions.first()?.first()?;
|
||||||
Some(Path(vec![*position]))
|
Some(Path(smallvec![*position]))
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO we must skip the successors that have already been sent
|
// TODO we must skip the successors that have already been sent
|
||||||
fn successors(&self, positions: &[Vec<u32>]) -> Vec<(Path, u32)> {
|
fn successors(&self, positions: &[Vec<u32>]) -> SmallVec16<(Path, u32)> {
|
||||||
let mut successors = Vec::new();
|
let mut successors = SmallVec16::new();
|
||||||
|
|
||||||
// If we can grow or shift the path
|
// If we can grow or shift the path
|
||||||
if self.0.len() < positions.len() {
|
if self.0.len() < positions.len() {
|
||||||
@ -103,7 +106,12 @@ impl Iterator for BestProximity {
|
|||||||
let result = dijkstra(
|
let result = dijkstra(
|
||||||
&Path::new(&self.positions)?,
|
&Path::new(&self.positions)?,
|
||||||
|p| p.successors(&self.positions),
|
|p| p.successors(&self.positions),
|
||||||
|p| self.is_path_successful(p) && output.as_ref().map_or(true, |(_, paths)| !paths.contains(&p.0)),
|
|p| {
|
||||||
|
self.is_path_successful(p) &&
|
||||||
|
output.as_ref().map_or(true, |(_, paths)| {
|
||||||
|
!paths.iter().position(|q| q.as_slice() == p.0.as_slice()).is_some()
|
||||||
|
})
|
||||||
|
},
|
||||||
);
|
);
|
||||||
|
|
||||||
match result {
|
match result {
|
||||||
@ -123,9 +131,9 @@ impl Iterator for BestProximity {
|
|||||||
|
|
||||||
// We add the new path to the output list as this path is known
|
// We add the new path to the output list as this path is known
|
||||||
// to be the requested distance.
|
// to be the requested distance.
|
||||||
paths.push(positions.0);
|
paths.push(positions.0.to_vec());
|
||||||
},
|
},
|
||||||
None => output = Some((positions.proximity(), vec![positions.0])),
|
None => output = Some((positions.proximity(), vec![positions.0.to_vec()])),
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
None => break,
|
None => break,
|
||||||
|
@ -45,10 +45,10 @@ struct Opt {
|
|||||||
|
|
||||||
struct Indexed {
|
struct Indexed {
|
||||||
fst: fst::Set<Vec<u8>>,
|
fst: fst::Set<Vec<u8>>,
|
||||||
postings_attrs: FastMap4<SmallVec32, RoaringBitmap>,
|
postings_attrs: FastMap4<SmallVec32<u8>, RoaringBitmap>,
|
||||||
prefix_postings_attrs: FastMap4<SmallVec32, RoaringBitmap>,
|
prefix_postings_attrs: FastMap4<SmallVec32<u8>, RoaringBitmap>,
|
||||||
postings_ids: FastMap4<SmallVec32, FastMap4<AttributeId, RoaringBitmap>>,
|
postings_ids: FastMap4<SmallVec32<u8>, FastMap4<AttributeId, RoaringBitmap>>,
|
||||||
prefix_postings_ids: FastMap4<SmallVec32, FastMap4<AttributeId, RoaringBitmap>>,
|
prefix_postings_ids: FastMap4<SmallVec32<u8>, FastMap4<AttributeId, RoaringBitmap>>,
|
||||||
headers: Vec<u8>,
|
headers: Vec<u8>,
|
||||||
documents: Vec<(DocumentId, Vec<u8>)>,
|
documents: Vec<(DocumentId, Vec<u8>)>,
|
||||||
}
|
}
|
||||||
|
@ -93,11 +93,11 @@ async fn main() -> anyhow::Result<()> {
|
|||||||
body.extend_from_slice(headers);
|
body.extend_from_slice(headers);
|
||||||
|
|
||||||
for id in documents_ids {
|
for id in documents_ids {
|
||||||
if let Some(content) = index.documents.get(&rtxn, &BEU32::new(id)).unwrap() {
|
let content = index.documents.get(&rtxn, &BEU32::new(id)).unwrap();
|
||||||
|
let content = content.expect(&format!("could not find document {}", id));
|
||||||
body.extend_from_slice(&content);
|
body.extend_from_slice(&content);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
Response::builder()
|
Response::builder()
|
||||||
.header("Content-Type", "text/csv")
|
.header("Content-Type", "text/csv")
|
||||||
|
59
src/lib.rs
59
src/lib.rs
@ -25,7 +25,8 @@ static LEVDIST2: Lazy<LevBuilder> = Lazy::new(|| LevBuilder::new(2, true));
|
|||||||
|
|
||||||
pub type FastMap4<K, V> = HashMap<K, V, BuildHasherDefault<FxHasher32>>;
|
pub type FastMap4<K, V> = HashMap<K, V, BuildHasherDefault<FxHasher32>>;
|
||||||
pub type SmallString32 = smallstr::SmallString<[u8; 32]>;
|
pub type SmallString32 = smallstr::SmallString<[u8; 32]>;
|
||||||
pub type SmallVec32 = smallvec::SmallVec<[u8; 32]>;
|
pub type SmallVec32<T> = smallvec::SmallVec<[T; 32]>;
|
||||||
|
pub type SmallVec16<T> = smallvec::SmallVec<[T; 16]>;
|
||||||
pub type BEU32 = heed::zerocopy::U32<heed::byteorder::BE>;
|
pub type BEU32 = heed::zerocopy::U32<heed::byteorder::BE>;
|
||||||
pub type DocumentId = u32;
|
pub type DocumentId = u32;
|
||||||
pub type AttributeId = u32;
|
pub type AttributeId = u32;
|
||||||
@ -89,22 +90,18 @@ impl Index {
|
|||||||
(word, is_prefix, dfa)
|
(word, is_prefix, dfa)
|
||||||
});
|
});
|
||||||
|
|
||||||
let mut words_positions = Vec::new();
|
let mut words = Vec::new();
|
||||||
let mut positions = Vec::new();
|
let mut positions = Vec::new();
|
||||||
let before = Instant::now();
|
let before = Instant::now();
|
||||||
|
|
||||||
for (word, is_prefix, dfa) in dfas {
|
for (word, _is_prefix, dfa) in dfas {
|
||||||
let mut count = 0;
|
let mut count = 0;
|
||||||
let mut union_positions = RoaringBitmap::default();
|
let mut union_positions = RoaringBitmap::default();
|
||||||
if false && word.len() <= 4 && is_prefix {
|
let mut derived_words = Vec::new();
|
||||||
if let Some(ids) = self.prefix_postings_attrs.get(rtxn, word.as_bytes())? {
|
// TODO re-enable the prefixes system
|
||||||
let right = RoaringBitmap::deserialize_from(ids)?;
|
|
||||||
union_positions.union_with(&right);
|
|
||||||
count = 1;
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
let mut stream = fst.search(&dfa).into_stream();
|
let mut stream = fst.search(&dfa).into_stream();
|
||||||
while let Some(word) = stream.next() {
|
while let Some(word) = stream.next() {
|
||||||
|
derived_words.push(word.to_vec());
|
||||||
let word = std::str::from_utf8(word)?;
|
let word = std::str::from_utf8(word)?;
|
||||||
if let Some(attrs) = self.postings_attrs.get(rtxn, word)? {
|
if let Some(attrs) = self.postings_attrs.get(rtxn, word)? {
|
||||||
let right = RoaringBitmap::deserialize_from(attrs)?;
|
let right = RoaringBitmap::deserialize_from(attrs)?;
|
||||||
@ -112,29 +109,22 @@ impl Index {
|
|||||||
count += 1;
|
count += 1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
eprintln!("{} words for {:?} we have found positions {:?}", count, word, union_positions);
|
eprintln!("{} words for {:?} we have found positions {:?}", count, word, union_positions);
|
||||||
words_positions.push((word, is_prefix, dfa));
|
words.push(derived_words);
|
||||||
positions.push(union_positions.iter().collect());
|
positions.push(union_positions.iter().collect());
|
||||||
}
|
}
|
||||||
|
|
||||||
eprintln!("Retrieving words positions took {:.02?}", before.elapsed());
|
eprintln!("Retrieving words positions took {:.02?}", before.elapsed());
|
||||||
|
|
||||||
// TODO re-enable the prefixes system
|
|
||||||
let mut words = Vec::new();
|
|
||||||
for (_word, _is_prefix, dfa) in words_positions {
|
|
||||||
let mut stream = fst.search(dfa).into_stream();
|
|
||||||
let mut derived_words = Vec::new();
|
|
||||||
while let Some(word) = stream.next() {
|
|
||||||
derived_words.push(word.to_vec());
|
|
||||||
}
|
|
||||||
words.push(derived_words);
|
|
||||||
}
|
|
||||||
|
|
||||||
let mut documents = Vec::new();
|
let mut documents = Vec::new();
|
||||||
|
|
||||||
'outer: for (proximity, positions) in BestProximity::new(positions) {
|
for (proximity, mut positions) in BestProximity::new(positions) {
|
||||||
|
// TODO we must ignore positions paths that gives nothing
|
||||||
|
if words.len() > 1 && proximity == 0 { continue }
|
||||||
|
|
||||||
|
positions.sort_unstable();
|
||||||
|
|
||||||
let same_prox_before = Instant::now();
|
let same_prox_before = Instant::now();
|
||||||
let mut same_proximity_union = RoaringBitmap::default();
|
let mut same_proximity_union = RoaringBitmap::default();
|
||||||
|
|
||||||
@ -177,15 +167,30 @@ impl Index {
|
|||||||
if let Some(intersect_docids) = intersect_docids {
|
if let Some(intersect_docids) = intersect_docids {
|
||||||
same_proximity_union.union_with(&intersect_docids);
|
same_proximity_union.union_with(&intersect_docids);
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
|
// We found enough documents we can stop here
|
||||||
|
if documents.iter().map(RoaringBitmap::len).sum::<u64>() + same_proximity_union.len() >= 20 {
|
||||||
eprintln!("proximity {} took a total of {:.02?}", proximity, same_prox_before.elapsed());
|
eprintln!("proximity {} took a total of {:.02?}", proximity, same_prox_before.elapsed());
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
documents.push(same_proximity_union);
|
documents.push(same_proximity_union);
|
||||||
|
|
||||||
// We found enough documents we can stop here
|
// We remove the double occurences of documents.
|
||||||
|
for i in 0..documents.len() {
|
||||||
|
if let Some((docs, others)) = documents[..=i].split_last_mut() {
|
||||||
|
others.iter().for_each(|other| docs.difference_with(other));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
documents.retain(|rb| !rb.is_empty());
|
||||||
|
|
||||||
|
eprintln!("documents: {:?}", documents);
|
||||||
|
eprintln!("proximity {} took a total of {:.02?}", proximity, same_prox_before.elapsed());
|
||||||
|
|
||||||
|
// We found enough documents we can stop here.
|
||||||
if documents.iter().map(RoaringBitmap::len).sum::<u64>() >= 20 {
|
if documents.iter().map(RoaringBitmap::len).sum::<u64>() >= 20 {
|
||||||
break 'outer;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user