mirror of
https://github.com/meilisearch/meilisearch.git
synced 2024-11-27 12:35:05 +08:00
527 lines
16 KiB
Rust
527 lines
16 KiB
Rust
use std::collections::BTreeMap;
|
|
use std::fs::File;
|
|
use std::io::{stdin, BufRead, BufReader, Cursor, Read, Write};
|
|
use std::path::PathBuf;
|
|
use std::str::FromStr;
|
|
use std::time::{Duration, Instant};
|
|
|
|
use byte_unit::Byte;
|
|
use eyre::Result;
|
|
use indicatif::{MultiProgress, ProgressBar, ProgressStyle};
|
|
use milli::documents::{DocumentsBatchBuilder, DocumentsBatchReader};
|
|
use milli::update::UpdateIndexingStep::{
|
|
ComputeIdsAndMergeDocuments, IndexDocuments, MergeDataIntoFinalDatabase, RemapDocumentAddition,
|
|
};
|
|
use milli::update::{self, IndexDocumentsConfig, IndexDocumentsMethod, IndexerConfig};
|
|
use milli::{heed, Index, Object};
|
|
use structopt::StructOpt;
|
|
|
|
#[global_allocator]
|
|
static ALLOC: mimalloc::MiMalloc = mimalloc::MiMalloc;
|
|
|
|
#[derive(Debug, StructOpt)]
|
|
#[structopt(name = "Milli CLI", about = "A simple CLI to manipulate a milli index.")]
|
|
struct Cli {
|
|
#[structopt(short, long, default_value = ".")]
|
|
index_path: PathBuf,
|
|
#[structopt(short = "s", long, default_value = "100GiB")]
|
|
index_size: Byte,
|
|
/// Verbose mode (-v, -vv, -vvv, etc.)
|
|
#[structopt(short, long, parse(from_occurrences))]
|
|
verbose: usize,
|
|
#[structopt(subcommand)]
|
|
subcommand: Command,
|
|
}
|
|
|
|
#[derive(Debug, StructOpt)]
|
|
enum Command {
|
|
Documents {
|
|
#[structopt(subcommand)]
|
|
cmd: Documents,
|
|
},
|
|
Search(Search),
|
|
Settings {
|
|
#[structopt(subcommand)]
|
|
cmd: Settings,
|
|
},
|
|
}
|
|
|
|
impl Performer for Command {
|
|
fn perform(self, index: Index) -> Result<()> {
|
|
match self {
|
|
Command::Documents { cmd } => cmd.perform(index),
|
|
Command::Search(cmd) => cmd.perform(index),
|
|
Command::Settings { cmd } => cmd.perform(index),
|
|
}
|
|
}
|
|
}
|
|
|
|
#[derive(Debug, StructOpt)]
|
|
enum Settings {
|
|
Update(SettingsUpdate),
|
|
Show,
|
|
}
|
|
|
|
impl Settings {
|
|
fn show(&self, index: Index) -> Result<()> {
|
|
let txn = index.read_txn()?;
|
|
let displayed_attributes = index
|
|
.displayed_fields(&txn)?
|
|
.map(|fields| fields.into_iter().map(String::from).collect());
|
|
|
|
let searchable_attributes: Option<Vec<_>> = index
|
|
.searchable_fields(&txn)?
|
|
.map(|fields| fields.into_iter().map(String::from).collect());
|
|
|
|
let filterable_attributes: Vec<_> = index.filterable_fields(&txn)?.into_iter().collect();
|
|
|
|
let sortable_attributes: Vec<_> = index.sortable_fields(&txn)?.into_iter().collect();
|
|
|
|
let criteria: Vec<_> = index.criteria(&txn)?.into_iter().map(|c| c.to_string()).collect();
|
|
|
|
let stop_words = index
|
|
.stop_words(&txn)?
|
|
.map(|stop_words| -> Result<Vec<_>> {
|
|
Ok(stop_words.stream().into_strs()?.into_iter().collect())
|
|
})
|
|
.transpose()?
|
|
.unwrap_or_else(Vec::new);
|
|
let distinct_field = index.distinct_field(&txn)?.map(String::from);
|
|
|
|
// in milli each word in the synonyms map were split on their separator. Since we lost
|
|
// this information we are going to put space between words.
|
|
let synonyms: BTreeMap<_, Vec<_>> = index
|
|
.synonyms(&txn)?
|
|
.iter()
|
|
.map(|(key, values)| {
|
|
(key.join(" "), values.iter().map(|value| value.join(" ")).collect())
|
|
})
|
|
.collect();
|
|
|
|
let exact_attributes = index.exact_attributes(&txn)?;
|
|
|
|
println!(
|
|
"displayed attributes:\n\t{}\nsearchable attributes:\n\t{}\nfilterable attributes:\n\t{}\nsortable attributes:\n\t{}\ncriterion:\n\t{}\nstop words:\n\t{}\ndistinct fields:\n\t{}\nsynonyms:\n\t{}\nexact attributes:\n\t{}\n",
|
|
displayed_attributes.unwrap_or(vec!["*".to_owned()]).join("\n\t"),
|
|
searchable_attributes.unwrap_or(vec!["*".to_owned()]).join("\n\t"),
|
|
filterable_attributes.join("\n\t"),
|
|
sortable_attributes.join("\n\t"),
|
|
criteria.join("\n\t"),
|
|
stop_words.join("\n\t"),
|
|
distinct_field.unwrap_or_default(),
|
|
synonyms.into_iter().map(|(k, v)| format!("\n\t{}:\n{:?}", k, v)).collect::<String>(),
|
|
exact_attributes.join("\n\t"),
|
|
);
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
impl Performer for Settings {
|
|
fn perform(self, index: Index) -> Result<()> {
|
|
match self {
|
|
Settings::Update(update) => update.perform(index),
|
|
Settings::Show => self.show(index),
|
|
}
|
|
}
|
|
}
|
|
|
|
#[derive(Debug, StructOpt)]
|
|
enum Documents {
|
|
Add(DocumentAddition),
|
|
}
|
|
|
|
impl Performer for Documents {
|
|
fn perform(self, index: Index) -> Result<()> {
|
|
match self {
|
|
Self::Add(addition) => addition.perform(index),
|
|
}
|
|
}
|
|
}
|
|
|
|
trait Performer {
|
|
fn perform(self, index: Index) -> Result<()>;
|
|
}
|
|
|
|
fn setup(opt: &Cli) -> Result<()> {
|
|
color_eyre::install()?;
|
|
stderrlog::new()
|
|
.verbosity(opt.verbose)
|
|
.show_level(false)
|
|
.timestamp(stderrlog::Timestamp::Off)
|
|
.init()?;
|
|
Ok(())
|
|
}
|
|
|
|
fn main() -> Result<()> {
|
|
let command = Cli::from_args();
|
|
|
|
setup(&command)?;
|
|
|
|
let mut options = heed::EnvOpenOptions::new();
|
|
options.map_size(command.index_size.get_bytes() as usize);
|
|
let index = milli::Index::new(options, command.index_path)?;
|
|
|
|
command.subcommand.perform(index)?;
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[derive(Debug)]
|
|
enum DocumentAdditionFormat {
|
|
Csv,
|
|
Json,
|
|
Jsonl,
|
|
}
|
|
|
|
impl FromStr for DocumentAdditionFormat {
|
|
type Err = eyre::Error;
|
|
|
|
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
|
match s {
|
|
"csv" => Ok(Self::Csv),
|
|
"jsonl" => Ok(Self::Jsonl),
|
|
"json" => Ok(Self::Json),
|
|
other => eyre::bail!("invalid format: {}", other),
|
|
}
|
|
}
|
|
}
|
|
|
|
#[derive(Debug, StructOpt)]
|
|
struct DocumentAddition {
|
|
#[structopt(short, long, default_value = "json", possible_values = &["csv", "jsonl", "json"])]
|
|
format: DocumentAdditionFormat,
|
|
/// Path to the update file, if not present, will read from stdin.
|
|
#[structopt(short, long)]
|
|
path: Option<PathBuf>,
|
|
/// Specify the primary key.
|
|
#[structopt(long)]
|
|
primary: Option<String>,
|
|
/// Whether to generate missing document ids.
|
|
#[structopt(short, long)]
|
|
autogen_docids: bool,
|
|
/// Whether to update or replace the documents if they already exist.
|
|
#[structopt(short, long)]
|
|
update_documents: bool,
|
|
}
|
|
|
|
impl Performer for DocumentAddition {
|
|
fn perform(self, index: milli::Index) -> Result<()> {
|
|
let reader: Box<dyn Read> = match self.path {
|
|
Some(ref path) => {
|
|
let file = File::open(path)?;
|
|
Box::new(file)
|
|
}
|
|
None => Box::new(stdin()),
|
|
};
|
|
|
|
println!("parsing documents...");
|
|
|
|
let reader = BufReader::new(reader);
|
|
|
|
let documents = match self.format {
|
|
DocumentAdditionFormat::Csv => documents_from_csv(reader)?,
|
|
DocumentAdditionFormat::Json => documents_from_json(reader)?,
|
|
DocumentAdditionFormat::Jsonl => documents_from_jsonl(reader)?,
|
|
};
|
|
|
|
let reader = DocumentsBatchReader::from_reader(Cursor::new(documents))?;
|
|
|
|
println!("Adding {} documents to the index.", reader.documents_count());
|
|
|
|
let mut txn = index.write_txn()?;
|
|
let config = milli::update::IndexerConfig { log_every_n: Some(100), ..Default::default() };
|
|
let update_method = if self.update_documents {
|
|
IndexDocumentsMethod::UpdateDocuments
|
|
} else {
|
|
IndexDocumentsMethod::ReplaceDocuments
|
|
};
|
|
|
|
if let Some(primary) = self.primary {
|
|
let mut builder = update::Settings::new(&mut txn, &index, &config);
|
|
builder.set_primary_key(primary);
|
|
builder.execute(|_| (), || false).unwrap();
|
|
}
|
|
|
|
let indexing_config = IndexDocumentsConfig {
|
|
update_method,
|
|
autogenerate_docids: self.autogen_docids,
|
|
..Default::default()
|
|
};
|
|
let mut bars = Vec::new();
|
|
let progesses = MultiProgress::new();
|
|
for _ in 0..4 {
|
|
let bar = ProgressBar::hidden();
|
|
let bar = progesses.add(bar);
|
|
bars.push(bar);
|
|
}
|
|
let addition = milli::update::IndexDocuments::new(
|
|
&mut txn,
|
|
&index,
|
|
&config,
|
|
indexing_config,
|
|
|step| indexing_callback(step, &bars),
|
|
|| false,
|
|
)
|
|
.unwrap();
|
|
let (addition, user_error) = addition.add_documents(reader)?;
|
|
if let Err(error) = user_error {
|
|
return Err(error.into());
|
|
}
|
|
|
|
let result = addition.execute()?;
|
|
|
|
txn.commit()?;
|
|
|
|
println!("{:?}", result);
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
fn indexing_callback(step: milli::update::UpdateIndexingStep, bars: &[ProgressBar]) {
|
|
let step_index = step.step();
|
|
let bar = &bars[step_index];
|
|
if step_index > 0 {
|
|
let prev = &bars[step_index - 1];
|
|
if !prev.is_finished() {
|
|
prev.disable_steady_tick();
|
|
prev.finish();
|
|
}
|
|
}
|
|
|
|
let style = ProgressStyle::default_bar()
|
|
.progress_chars("##-")
|
|
.template("[eta: {eta_precise}] {bar:40.cyan/blue} {pos:>7}/{len:7} {msg}")
|
|
.unwrap();
|
|
|
|
match step {
|
|
RemapDocumentAddition { documents_seen } => {
|
|
bar.set_style(ProgressStyle::default_spinner());
|
|
bar.set_message(format!("remapped {} documents so far.", documents_seen));
|
|
}
|
|
ComputeIdsAndMergeDocuments { documents_seen, total_documents } => {
|
|
bar.set_style(style);
|
|
bar.set_length(total_documents as u64);
|
|
bar.set_message("Merging documents...");
|
|
bar.set_position(documents_seen as u64);
|
|
}
|
|
IndexDocuments { documents_seen, total_documents } => {
|
|
bar.set_style(style);
|
|
bar.set_length(total_documents as u64);
|
|
bar.set_message("Indexing documents...");
|
|
bar.set_position(documents_seen as u64);
|
|
}
|
|
MergeDataIntoFinalDatabase { databases_seen, total_databases } => {
|
|
bar.set_style(style);
|
|
bar.set_length(total_databases as u64);
|
|
bar.set_message("Merging databases...");
|
|
bar.set_position(databases_seen as u64);
|
|
}
|
|
}
|
|
bar.enable_steady_tick(Duration::from_millis(200));
|
|
}
|
|
|
|
fn documents_from_jsonl(reader: impl Read) -> Result<Vec<u8>> {
|
|
let mut documents = DocumentsBatchBuilder::new(Vec::new());
|
|
let reader = BufReader::new(reader);
|
|
|
|
for result in serde_json::Deserializer::from_reader(reader).into_iter::<Object>() {
|
|
let object = result?;
|
|
documents.append_json_object(&object)?;
|
|
}
|
|
|
|
documents.into_inner().map_err(Into::into)
|
|
}
|
|
|
|
fn documents_from_json(reader: impl Read) -> Result<Vec<u8>> {
|
|
let mut documents = DocumentsBatchBuilder::new(Vec::new());
|
|
|
|
documents.append_json_array(reader)?;
|
|
|
|
documents.into_inner().map_err(Into::into)
|
|
}
|
|
|
|
fn documents_from_csv(reader: impl Read) -> Result<Vec<u8>> {
|
|
let csv = csv::Reader::from_reader(reader);
|
|
|
|
let mut documents = DocumentsBatchBuilder::new(Vec::new());
|
|
documents.append_csv(csv)?;
|
|
|
|
documents.into_inner().map_err(Into::into)
|
|
}
|
|
|
|
#[derive(Debug, StructOpt)]
|
|
struct Search {
|
|
query: Option<String>,
|
|
#[structopt(short, long)]
|
|
filter: Option<String>,
|
|
#[structopt(short, long)]
|
|
offset: Option<usize>,
|
|
#[structopt(short, long)]
|
|
limit: Option<usize>,
|
|
#[structopt(short, long, conflicts_with = "query")]
|
|
interactive: bool,
|
|
}
|
|
|
|
impl Performer for Search {
|
|
fn perform(self, index: milli::Index) -> Result<()> {
|
|
if self.interactive {
|
|
let stdin = std::io::stdin();
|
|
let mut lines = stdin.lock().lines();
|
|
loop {
|
|
eprint!("> ");
|
|
std::io::stdout().flush()?;
|
|
match lines.next() {
|
|
Some(Ok(line)) => {
|
|
let now = Instant::now();
|
|
let jsons = Self::perform_single_search(
|
|
&index,
|
|
&Some(line),
|
|
&self.filter,
|
|
&self.offset,
|
|
&self.limit,
|
|
)?;
|
|
|
|
let time = now.elapsed();
|
|
|
|
let hits = serde_json::to_string_pretty(&jsons)?;
|
|
|
|
println!("{}", hits);
|
|
eprintln!("found {} results in {:.02?}", jsons.len(), time);
|
|
}
|
|
_ => break,
|
|
}
|
|
}
|
|
} else {
|
|
let now = Instant::now();
|
|
let jsons = Self::perform_single_search(
|
|
&index,
|
|
&self.query,
|
|
&self.filter,
|
|
&self.offset,
|
|
&self.limit,
|
|
)?;
|
|
|
|
let time = now.elapsed();
|
|
|
|
let hits = serde_json::to_string_pretty(&jsons)?;
|
|
|
|
println!("{}", hits);
|
|
eprintln!("found {} results in {:.02?}", jsons.len(), time);
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
impl Search {
|
|
fn perform_single_search(
|
|
index: &milli::Index,
|
|
query: &Option<String>,
|
|
filter: &Option<String>,
|
|
offset: &Option<usize>,
|
|
limit: &Option<usize>,
|
|
) -> Result<Vec<Object>> {
|
|
let txn = index.read_txn()?;
|
|
let mut search = index.search(&txn);
|
|
|
|
if let Some(ref query) = query {
|
|
search.query(query);
|
|
}
|
|
|
|
if let Some(ref filter) = filter {
|
|
if let Some(condition) = milli::Filter::from_str(filter)? {
|
|
search.filter(condition);
|
|
}
|
|
}
|
|
|
|
if let Some(offset) = offset {
|
|
search.offset(*offset);
|
|
}
|
|
|
|
if let Some(limit) = limit {
|
|
search.limit(*limit);
|
|
}
|
|
|
|
let result = search.execute()?;
|
|
|
|
let fields_ids_map = index.fields_ids_map(&txn)?;
|
|
let displayed_fields =
|
|
index.displayed_fields_ids(&txn)?.unwrap_or_else(|| fields_ids_map.ids().collect());
|
|
let documents = index.documents(&txn, result.documents_ids)?;
|
|
let mut jsons = Vec::new();
|
|
for (_, obkv) in documents {
|
|
let json = milli::obkv_to_json(&displayed_fields, &fields_ids_map, obkv)?;
|
|
jsons.push(json);
|
|
}
|
|
|
|
Ok(jsons)
|
|
}
|
|
}
|
|
|
|
#[derive(Debug, StructOpt)]
|
|
struct SettingsUpdate {
|
|
#[structopt(long)]
|
|
filterable_attributes: Option<Vec<String>>,
|
|
#[structopt(long)]
|
|
criteria: Option<Vec<String>>,
|
|
#[structopt(long)]
|
|
exact_attributes: Option<Vec<String>>,
|
|
#[structopt(long)]
|
|
distinct_attribute: Option<String>,
|
|
}
|
|
|
|
impl Performer for SettingsUpdate {
|
|
fn perform(self, index: milli::Index) -> Result<()> {
|
|
let mut txn = index.write_txn()?;
|
|
|
|
let config = IndexerConfig { log_every_n: Some(100), ..Default::default() };
|
|
|
|
let mut update = milli::update::Settings::new(&mut txn, &index, &config);
|
|
|
|
if let Some(ref filterable_attributes) = self.filterable_attributes {
|
|
if !filterable_attributes.is_empty() {
|
|
update.set_filterable_fields(filterable_attributes.iter().cloned().collect());
|
|
} else {
|
|
update.reset_filterable_fields();
|
|
}
|
|
}
|
|
|
|
if let Some(criteria) = self.criteria {
|
|
if !criteria.is_empty() {
|
|
update.set_criteria(criteria);
|
|
} else {
|
|
update.reset_criteria();
|
|
}
|
|
}
|
|
|
|
if let Some(exact_attributes) = self.exact_attributes {
|
|
if !exact_attributes.is_empty() {
|
|
update.set_exact_attributes(exact_attributes.into_iter().collect());
|
|
} else {
|
|
update.reset_exact_attributes();
|
|
}
|
|
}
|
|
|
|
if let Some(distinct_attr) = self.distinct_attribute {
|
|
if !distinct_attr.is_empty() {
|
|
update.set_distinct_field(distinct_attr);
|
|
} else {
|
|
update.reset_distinct_field();
|
|
}
|
|
}
|
|
|
|
let mut bars = Vec::new();
|
|
let progesses = MultiProgress::new();
|
|
for _ in 0..4 {
|
|
let bar = ProgressBar::hidden();
|
|
let bar = progesses.add(bar);
|
|
bars.push(bar);
|
|
}
|
|
|
|
update.execute(|step| indexing_callback(step, &bars), || false)?;
|
|
|
|
txn.commit()?;
|
|
Ok(())
|
|
}
|
|
}
|