2020-08-17 21:15:37 +08:00
|
|
|
use std::path::PathBuf;
|
|
|
|
use std::{str, io};
|
|
|
|
|
2020-08-23 16:52:47 +08:00
|
|
|
use anyhow::Context;
|
2020-08-17 21:15:37 +08:00
|
|
|
use heed::EnvOpenOptions;
|
|
|
|
use milli::Index;
|
|
|
|
use structopt::StructOpt;
|
|
|
|
|
2020-08-23 16:52:47 +08:00
|
|
|
use Command::*;
|
|
|
|
|
2020-08-17 21:15:37 +08:00
|
|
|
#[cfg(target_os = "linux")]
|
|
|
|
#[global_allocator]
|
|
|
|
static ALLOC: jemallocator::Jemalloc = jemallocator::Jemalloc;
|
|
|
|
|
|
|
|
#[derive(Debug, StructOpt)]
|
|
|
|
#[structopt(name = "milli-info", about = "A stats crawler for milli.")]
|
|
|
|
struct Opt {
|
|
|
|
/// The database path where the database is located.
|
|
|
|
/// It is created if it doesn't already exist.
|
|
|
|
#[structopt(long = "db", parse(from_os_str))]
|
|
|
|
database: PathBuf,
|
|
|
|
|
|
|
|
/// The maximum size the database can take on disk. It is recommended to specify
|
|
|
|
/// the whole disk space (value must be a multiple of a page size).
|
|
|
|
#[structopt(long = "db-size", default_value = "107374182400")] // 100 GB
|
|
|
|
database_size: usize,
|
|
|
|
|
|
|
|
/// Verbose mode (-v, -vv, -vvv, etc.)
|
|
|
|
#[structopt(short, long, parse(from_occurrences))]
|
|
|
|
verbose: usize,
|
|
|
|
|
|
|
|
#[structopt(subcommand)]
|
|
|
|
command: Command,
|
|
|
|
}
|
|
|
|
|
|
|
|
#[derive(Debug, StructOpt)]
|
|
|
|
enum Command {
|
|
|
|
/// Outputs a CSV of the most frequent words of this index.
|
|
|
|
///
|
|
|
|
/// `word` are displayed and ordered by frequency.
|
|
|
|
/// `document_frequency` defines the number of documents which contains the word.
|
|
|
|
MostCommonWords {
|
|
|
|
/// The maximum number of frequencies to return.
|
|
|
|
#[structopt(default_value = "10")]
|
|
|
|
limit: usize,
|
2020-08-21 20:24:05 +08:00
|
|
|
},
|
|
|
|
|
2020-08-21 20:42:55 +08:00
|
|
|
/// Outputs a CSV with the biggest entries of the database.
|
2020-09-06 23:14:20 +08:00
|
|
|
BiggestValues {
|
2020-08-21 20:42:55 +08:00
|
|
|
/// The maximum number of sizes to return.
|
|
|
|
#[structopt(default_value = "10")]
|
|
|
|
limit: usize,
|
|
|
|
},
|
2020-08-23 16:52:47 +08:00
|
|
|
|
2020-09-07 20:56:48 +08:00
|
|
|
/// Outputs the total size of all the docid-word-positions keys and values.
|
|
|
|
TotalDocidWordPositionsSize,
|
|
|
|
|
|
|
|
/// Outputs the average number of *different* words by document.
|
|
|
|
AverageNumberOfWordsByDoc,
|
|
|
|
|
2020-09-07 21:26:42 +08:00
|
|
|
/// Outputs the average number of positions for each document words.
|
|
|
|
AverageNumberOfPositions,
|
|
|
|
|
2020-09-06 23:14:20 +08:00
|
|
|
/// Outputs the words FST to disk.
|
|
|
|
///
|
|
|
|
/// One can use the FST binary helper to dissect and analyze it,
|
|
|
|
/// you can install it using `cargo install fst-bin`.
|
|
|
|
ExportWordsFst {
|
|
|
|
/// The path where the FST will be written.
|
|
|
|
#[structopt(short, long, default_value = "words.fst")]
|
|
|
|
output: PathBuf,
|
2020-08-30 00:12:31 +08:00
|
|
|
},
|
2020-08-17 21:15:37 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
fn main() -> anyhow::Result<()> {
|
|
|
|
let opt = Opt::from_args();
|
|
|
|
|
|
|
|
stderrlog::new()
|
|
|
|
.verbosity(opt.verbose)
|
|
|
|
.show_level(false)
|
|
|
|
.timestamp(stderrlog::Timestamp::Off)
|
|
|
|
.init()?;
|
|
|
|
|
|
|
|
let env = EnvOpenOptions::new()
|
|
|
|
.map_size(opt.database_size)
|
|
|
|
.max_dbs(10)
|
|
|
|
.open(&opt.database)?;
|
|
|
|
|
|
|
|
// Open the LMDB database.
|
2020-08-28 21:38:05 +08:00
|
|
|
let index = Index::new(&env)?;
|
2020-08-17 21:15:37 +08:00
|
|
|
let rtxn = env.read_txn()?;
|
|
|
|
|
|
|
|
match opt.command {
|
2020-08-23 16:52:47 +08:00
|
|
|
MostCommonWords { limit } => most_common_words(&index, &rtxn, limit),
|
2020-09-06 23:14:20 +08:00
|
|
|
BiggestValues { limit } => biggest_value_sizes(&index, &rtxn, limit),
|
2020-09-07 20:56:48 +08:00
|
|
|
TotalDocidWordPositionsSize => total_docid_word_positions_size(&index, &rtxn),
|
|
|
|
AverageNumberOfWordsByDoc => average_number_of_words_by_doc(&index, &rtxn),
|
2020-09-07 21:26:42 +08:00
|
|
|
AverageNumberOfPositions => average_number_of_positions(&index, &rtxn),
|
2020-09-06 23:14:20 +08:00
|
|
|
ExportWordsFst { output } => export_words_fst(&index, &rtxn, output),
|
2020-08-17 21:15:37 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
fn most_common_words(index: &Index, rtxn: &heed::RoTxn, limit: usize) -> anyhow::Result<()> {
|
|
|
|
use std::collections::BinaryHeap;
|
|
|
|
use std::cmp::Reverse;
|
|
|
|
|
|
|
|
let mut heap = BinaryHeap::with_capacity(limit + 1);
|
2020-09-06 23:14:20 +08:00
|
|
|
for result in index.word_docids.iter(rtxn)? {
|
2020-08-17 21:15:37 +08:00
|
|
|
if limit == 0 { break }
|
2020-09-06 23:14:20 +08:00
|
|
|
let (word, docids) = result?;
|
|
|
|
heap.push((Reverse(docids.len()), word));
|
2020-08-17 21:15:37 +08:00
|
|
|
if heap.len() > limit { heap.pop(); }
|
|
|
|
}
|
|
|
|
|
|
|
|
let stdout = io::stdout();
|
|
|
|
let mut wtr = csv::Writer::from_writer(stdout.lock());
|
2020-09-06 23:14:20 +08:00
|
|
|
wtr.write_record(&["word", "document_frequency"])?;
|
2020-08-17 21:15:37 +08:00
|
|
|
|
2020-09-06 23:14:20 +08:00
|
|
|
for (Reverse(document_frequency), word) in heap.into_sorted_vec() {
|
|
|
|
wtr.write_record(&[word, &document_frequency.to_string()])?;
|
2020-08-21 20:24:05 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
Ok(wtr.flush()?)
|
|
|
|
}
|
2020-08-21 20:42:55 +08:00
|
|
|
|
|
|
|
fn biggest_value_sizes(index: &Index, rtxn: &heed::RoTxn, limit: usize) -> anyhow::Result<()> {
|
|
|
|
use std::cmp::Reverse;
|
|
|
|
use std::collections::BinaryHeap;
|
|
|
|
use heed::types::{Str, ByteSlice};
|
2020-09-06 23:14:20 +08:00
|
|
|
use milli::heed_codec::BEU32StrCodec;
|
2020-08-21 20:42:55 +08:00
|
|
|
|
2020-08-26 20:36:22 +08:00
|
|
|
let main_name = "main";
|
2020-09-06 23:14:20 +08:00
|
|
|
let word_docids_name = "word_docids";
|
|
|
|
let docid_word_positions_name = "docid_word_positions";
|
2020-08-21 20:42:55 +08:00
|
|
|
|
|
|
|
let mut heap = BinaryHeap::with_capacity(limit + 1);
|
|
|
|
|
|
|
|
if limit > 0 {
|
2020-08-26 20:36:22 +08:00
|
|
|
if let Some(fst) = index.fst(rtxn)? {
|
|
|
|
heap.push(Reverse((fst.as_fst().as_bytes().len(), format!("words-fst"), main_name)));
|
|
|
|
if heap.len() > limit { heap.pop(); }
|
|
|
|
}
|
|
|
|
|
2020-08-30 00:12:31 +08:00
|
|
|
if let Some(documents) = index.main.get::<_, Str, ByteSlice>(rtxn, "documents")? {
|
2020-08-28 21:38:05 +08:00
|
|
|
heap.push(Reverse((documents.len(), format!("documents"), main_name)));
|
|
|
|
if heap.len() > limit { heap.pop(); }
|
|
|
|
}
|
|
|
|
|
2020-08-30 00:12:31 +08:00
|
|
|
if let Some(documents_ids) = index.main.get::<_, Str, ByteSlice>(rtxn, "documents-ids")? {
|
|
|
|
heap.push(Reverse((documents_ids.len(), format!("documents-ids"), main_name)));
|
|
|
|
if heap.len() > limit { heap.pop(); }
|
|
|
|
}
|
|
|
|
|
2020-09-06 23:14:20 +08:00
|
|
|
for result in index.word_docids.as_polymorph().iter::<_, Str, ByteSlice>(rtxn)? {
|
2020-08-21 20:42:55 +08:00
|
|
|
let (word, value) = result?;
|
2020-09-06 23:14:20 +08:00
|
|
|
heap.push(Reverse((value.len(), word.to_string(), word_docids_name)));
|
2020-08-21 20:42:55 +08:00
|
|
|
if heap.len() > limit { heap.pop(); }
|
|
|
|
}
|
|
|
|
|
2020-09-06 23:14:20 +08:00
|
|
|
for result in index.docid_word_positions.as_polymorph().iter::<_, BEU32StrCodec, ByteSlice>(rtxn)? {
|
|
|
|
let ((docid, word), value) = result?;
|
|
|
|
let key = format!("{} {}", docid, word);
|
|
|
|
heap.push(Reverse((value.len(), key, docid_word_positions_name)));
|
2020-08-21 20:42:55 +08:00
|
|
|
if heap.len() > limit { heap.pop(); }
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
let stdout = io::stdout();
|
|
|
|
let mut wtr = csv::Writer::from_writer(stdout.lock());
|
2020-08-26 20:36:22 +08:00
|
|
|
wtr.write_record(&["database_name", "key_name", "size"])?;
|
2020-08-21 20:42:55 +08:00
|
|
|
|
|
|
|
for Reverse((size, key_name, database_name)) in heap.into_sorted_vec() {
|
|
|
|
wtr.write_record(&[database_name.to_string(), key_name, size.to_string()])?;
|
|
|
|
}
|
|
|
|
|
|
|
|
Ok(wtr.flush()?)
|
2020-08-23 16:52:47 +08:00
|
|
|
}
|
|
|
|
|
2020-09-06 23:14:20 +08:00
|
|
|
fn export_words_fst(index: &Index, rtxn: &heed::RoTxn, output: PathBuf) -> anyhow::Result<()> {
|
|
|
|
use std::fs::File;
|
|
|
|
use std::io::Write as _;
|
2020-08-23 16:52:47 +08:00
|
|
|
|
2020-09-06 23:14:20 +08:00
|
|
|
let mut output = File::create(&output)
|
|
|
|
.with_context(|| format!("failed to create {} file", output.display()))?;
|
2020-08-23 16:52:47 +08:00
|
|
|
|
2020-09-06 23:14:20 +08:00
|
|
|
match index.fst(rtxn)? {
|
|
|
|
Some(fst) => output.write_all(fst.as_fst().as_bytes())?,
|
|
|
|
None => {
|
|
|
|
let fst = fst::Set::default();
|
|
|
|
output.write_all(fst.as_fst().as_bytes())?;
|
|
|
|
},
|
2020-08-23 16:52:47 +08:00
|
|
|
}
|
|
|
|
|
2020-09-06 23:14:20 +08:00
|
|
|
Ok(())
|
2020-08-21 20:42:55 +08:00
|
|
|
}
|
2020-09-07 20:56:48 +08:00
|
|
|
|
|
|
|
fn total_docid_word_positions_size(index: &Index, rtxn: &heed::RoTxn) -> anyhow::Result<()> {
|
|
|
|
use heed::types::ByteSlice;
|
|
|
|
|
|
|
|
let mut total_key_size = 0;
|
|
|
|
let mut total_val_size = 0;
|
|
|
|
let mut count = 0;
|
|
|
|
|
|
|
|
let iter = index.docid_word_positions.as_polymorph().iter::<_, ByteSlice, ByteSlice>(rtxn)?;
|
|
|
|
for result in iter {
|
|
|
|
let (key, val) = result?;
|
|
|
|
total_key_size += key.len();
|
|
|
|
total_val_size += val.len();
|
|
|
|
count += 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
println!("number of keys: {}", count);
|
|
|
|
println!("total key size: {}", total_key_size);
|
|
|
|
println!("total value size: {}", total_val_size);
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
|
|
|
fn average_number_of_words_by_doc(index: &Index, rtxn: &heed::RoTxn) -> anyhow::Result<()> {
|
|
|
|
use heed::types::DecodeIgnore;
|
|
|
|
use milli::{DocumentId, BEU32StrCodec};
|
|
|
|
|
|
|
|
let mut words_counts = Vec::new();
|
|
|
|
let mut count = 0;
|
|
|
|
let mut prev = None as Option<(DocumentId, u32)>;
|
|
|
|
|
|
|
|
let iter = index.docid_word_positions.as_polymorph().iter::<_, BEU32StrCodec, DecodeIgnore>(rtxn)?;
|
|
|
|
for result in iter {
|
|
|
|
let ((docid, _word), ()) = result?;
|
|
|
|
|
|
|
|
match prev.as_mut() {
|
|
|
|
Some((prev_docid, prev_count)) if docid == *prev_docid => {
|
|
|
|
*prev_count += 1;
|
|
|
|
},
|
|
|
|
Some((prev_docid, prev_count)) => {
|
|
|
|
words_counts.push(*prev_count);
|
|
|
|
*prev_docid = docid;
|
|
|
|
*prev_count = 0;
|
|
|
|
count += 1;
|
|
|
|
},
|
|
|
|
None => prev = Some((docid, 1)),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if let Some((_, prev_count)) = prev.take() {
|
|
|
|
words_counts.push(prev_count);
|
|
|
|
count += 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
let words_count = words_counts.into_iter().map(|c| c as usize).sum::<usize>() as f64;
|
|
|
|
let count = count as f64;
|
|
|
|
|
|
|
|
println!("average number of different words by document: {}", words_count / count);
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
2020-09-07 21:26:42 +08:00
|
|
|
|
|
|
|
fn average_number_of_positions(index: &Index, rtxn: &heed::RoTxn) -> anyhow::Result<()> {
|
|
|
|
use heed::types::DecodeIgnore;
|
2020-09-07 21:42:20 +08:00
|
|
|
use milli::ByteorderXRoaringBitmapCodec;
|
2020-09-07 21:26:42 +08:00
|
|
|
|
|
|
|
let mut values_length = Vec::new();
|
|
|
|
let mut count = 0;
|
|
|
|
|
2020-09-07 21:42:20 +08:00
|
|
|
let db = index.docid_word_positions.as_polymorph();
|
|
|
|
for result in db.iter::<_, DecodeIgnore, ByteorderXRoaringBitmapCodec>(rtxn)? {
|
2020-09-07 21:26:42 +08:00
|
|
|
let ((), val) = result?;
|
|
|
|
values_length.push(val.len() as u32);
|
|
|
|
count += 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
let values_length_sum = values_length.into_iter().map(|c| c as usize).sum::<usize>() as f64;
|
|
|
|
let count = count as f64;
|
|
|
|
|
|
|
|
println!("average number of positions by word: {}", values_length_sum / count);
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|