mirror of
https://github.com/meilisearch/meilisearch.git
synced 2024-11-23 10:37:41 +08:00
Move the heed env into the index itself to ease the usage of the library
This commit is contained in:
parent
b5d52b6b45
commit
e63fdf2b22
32
src/index.rs
32
src/index.rs
@ -1,4 +1,5 @@
|
|||||||
use std::borrow::Cow;
|
use std::borrow::Cow;
|
||||||
|
use std::path::Path;
|
||||||
|
|
||||||
use anyhow::Context;
|
use anyhow::Context;
|
||||||
use heed::types::*;
|
use heed::types::*;
|
||||||
@ -20,6 +21,8 @@ pub const USERS_IDS_DOCUMENTS_IDS_KEY: &str = "users-ids-documents-ids";
|
|||||||
|
|
||||||
#[derive(Clone)]
|
#[derive(Clone)]
|
||||||
pub struct Index {
|
pub struct Index {
|
||||||
|
/// The LMDB environment which this index is associated with.
|
||||||
|
pub env: heed::Env,
|
||||||
/// Contains many different types (e.g. the fields ids map).
|
/// Contains many different types (e.g. the fields ids map).
|
||||||
pub main: PolyDatabase,
|
pub main: PolyDatabase,
|
||||||
/// A word and all the documents ids containing the word.
|
/// A word and all the documents ids containing the word.
|
||||||
@ -33,14 +36,27 @@ pub struct Index {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl Index {
|
impl Index {
|
||||||
pub fn new(env: &heed::Env) -> anyhow::Result<Index> {
|
pub fn new<P: AsRef<Path>>(mut options: heed::EnvOpenOptions, path: P) -> anyhow::Result<Index> {
|
||||||
Ok(Index {
|
options.max_dbs(5);
|
||||||
main: env.create_poly_database(Some("main"))?,
|
|
||||||
word_docids: env.create_database(Some("word-docids"))?,
|
let env = options.open(path)?;
|
||||||
docid_word_positions: env.create_database(Some("docid-word-positions"))?,
|
let main = env.create_poly_database(Some("main"))?;
|
||||||
word_pair_proximity_docids: env.create_database(Some("word-pair-proximity-docids"))?,
|
let word_docids = env.create_database(Some("word-docids"))?;
|
||||||
documents: env.create_database(Some("documents"))?,
|
let docid_word_positions = env.create_database(Some("docid-word-positions"))?;
|
||||||
})
|
let word_pair_proximity_docids = env.create_database(Some("word-pair-proximity-docids"))?;
|
||||||
|
let documents = env.create_database(Some("documents"))?;
|
||||||
|
|
||||||
|
Ok(Index { env, main, word_docids, docid_word_positions, word_pair_proximity_docids, documents })
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a write transaction to be able to write into the index.
|
||||||
|
pub fn write_txn(&self) -> heed::Result<heed::RwTxn> {
|
||||||
|
self.env.write_txn()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a read transaction to be able to read the index.
|
||||||
|
pub fn read_txn(&self) -> heed::Result<heed::RoTxn> {
|
||||||
|
self.env.read_txn()
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Writes the documents ids that corresponds to the user-ids-documents-ids FST.
|
/// Writes the documents ids that corresponds to the user-ids-documents-ids FST.
|
||||||
|
@ -130,20 +130,18 @@ enum Command {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn run(opt: Opt) -> anyhow::Result<()> {
|
pub fn run(opt: Opt) -> anyhow::Result<()> {
|
||||||
let env = EnvOpenOptions::new()
|
|
||||||
.map_size(opt.database_size)
|
|
||||||
.max_dbs(10)
|
|
||||||
.open(&opt.database)?;
|
|
||||||
|
|
||||||
stderrlog::new()
|
stderrlog::new()
|
||||||
.verbosity(opt.verbose)
|
.verbosity(opt.verbose)
|
||||||
.show_level(false)
|
.show_level(false)
|
||||||
.timestamp(stderrlog::Timestamp::Off)
|
.timestamp(stderrlog::Timestamp::Off)
|
||||||
.init()?;
|
.init()?;
|
||||||
|
|
||||||
|
let mut options = EnvOpenOptions::new();
|
||||||
|
options.map_size(opt.database_size);
|
||||||
|
|
||||||
// Open the LMDB database.
|
// Open the LMDB database.
|
||||||
let index = Index::new(&env)?;
|
let index = Index::new(options, opt.database)?;
|
||||||
let rtxn = env.read_txn()?;
|
let rtxn = index.read_txn()?;
|
||||||
|
|
||||||
match opt.command {
|
match opt.command {
|
||||||
MostCommonWords { limit } => most_common_words(&index, &rtxn, limit),
|
MostCommonWords { limit } => most_common_words(&index, &rtxn, limit),
|
||||||
|
@ -40,14 +40,12 @@ pub fn run(opt: Opt) -> anyhow::Result<()> {
|
|||||||
.init()?;
|
.init()?;
|
||||||
|
|
||||||
std::fs::create_dir_all(&opt.database)?;
|
std::fs::create_dir_all(&opt.database)?;
|
||||||
let env = EnvOpenOptions::new()
|
let mut options = EnvOpenOptions::new();
|
||||||
.map_size(opt.database_size)
|
options.map_size(opt.database_size);
|
||||||
.max_dbs(10)
|
|
||||||
.open(&opt.database)?;
|
|
||||||
|
|
||||||
// Open the LMDB database.
|
// Open the LMDB database.
|
||||||
let index = Index::new(&env)?;
|
let index = Index::new(options, &opt.database)?;
|
||||||
let rtxn = env.read_txn()?;
|
let rtxn = index.read_txn()?;
|
||||||
|
|
||||||
let stdin = io::stdin();
|
let stdin = io::stdin();
|
||||||
let lines = match opt.query {
|
let lines = match opt.query {
|
||||||
|
@ -181,13 +181,11 @@ pub fn run(opt: Opt) -> anyhow::Result<()> {
|
|||||||
.init()?;
|
.init()?;
|
||||||
|
|
||||||
create_dir_all(&opt.database)?;
|
create_dir_all(&opt.database)?;
|
||||||
let env = EnvOpenOptions::new()
|
let mut options = EnvOpenOptions::new();
|
||||||
.map_size(opt.database_size)
|
options.map_size(opt.database_size);
|
||||||
.max_dbs(10)
|
|
||||||
.open(&opt.database)?;
|
|
||||||
|
|
||||||
// Open the LMDB database.
|
// Open the LMDB database.
|
||||||
let index = Index::new(&env)?;
|
let index = Index::new(options, &opt.database)?;
|
||||||
|
|
||||||
// Setup the LMDB based update database.
|
// Setup the LMDB based update database.
|
||||||
let mut update_store_options = EnvOpenOptions::new();
|
let mut update_store_options = EnvOpenOptions::new();
|
||||||
@ -198,7 +196,6 @@ pub fn run(opt: Opt) -> anyhow::Result<()> {
|
|||||||
|
|
||||||
let (update_status_sender, _) = broadcast::channel(100);
|
let (update_status_sender, _) = broadcast::channel(100);
|
||||||
let update_status_sender_cloned = update_status_sender.clone();
|
let update_status_sender_cloned = update_status_sender.clone();
|
||||||
let env_cloned = env.clone();
|
|
||||||
let index_cloned = index.clone();
|
let index_cloned = index.clone();
|
||||||
let indexer_opt_cloned = opt.indexer.clone();
|
let indexer_opt_cloned = opt.indexer.clone();
|
||||||
let update_store = UpdateStore::open(
|
let update_store = UpdateStore::open(
|
||||||
@ -226,7 +223,7 @@ pub fn run(opt: Opt) -> anyhow::Result<()> {
|
|||||||
let result: anyhow::Result<()> = match meta {
|
let result: anyhow::Result<()> = match meta {
|
||||||
UpdateMeta::DocumentsAddition => {
|
UpdateMeta::DocumentsAddition => {
|
||||||
// We must use the write transaction of the update here.
|
// We must use the write transaction of the update here.
|
||||||
let mut wtxn = env_cloned.write_txn()?;
|
let mut wtxn = index_cloned.write_txn()?;
|
||||||
let mut builder = update_builder.index_documents(&mut wtxn, &index_cloned);
|
let mut builder = update_builder.index_documents(&mut wtxn, &index_cloned);
|
||||||
|
|
||||||
let replace_documents = true;
|
let replace_documents = true;
|
||||||
@ -283,7 +280,6 @@ pub fn run(opt: Opt) -> anyhow::Result<()> {
|
|||||||
// Expose an HTML page to debug the search in a browser
|
// Expose an HTML page to debug the search in a browser
|
||||||
let db_name_cloned = db_name.clone();
|
let db_name_cloned = db_name.clone();
|
||||||
let lmdb_path_cloned = lmdb_path.clone();
|
let lmdb_path_cloned = lmdb_path.clone();
|
||||||
let env_cloned = env.clone();
|
|
||||||
let index_cloned = index.clone();
|
let index_cloned = index.clone();
|
||||||
let dash_html_route = warp::filters::method::get()
|
let dash_html_route = warp::filters::method::get()
|
||||||
.and(warp::filters::path::end())
|
.and(warp::filters::path::end())
|
||||||
@ -296,7 +292,7 @@ pub fn run(opt: Opt) -> anyhow::Result<()> {
|
|||||||
.len() as usize;
|
.len() as usize;
|
||||||
|
|
||||||
// And the number of documents in the database.
|
// And the number of documents in the database.
|
||||||
let rtxn = env_cloned.clone().read_txn().unwrap();
|
let rtxn = index_cloned.clone().read_txn().unwrap();
|
||||||
let docs_count = index_cloned.clone().number_of_documents(&rtxn).unwrap() as usize;
|
let docs_count = index_cloned.clone().number_of_documents(&rtxn).unwrap() as usize;
|
||||||
|
|
||||||
IndexTemplate { db_name: db_name_cloned.clone(), db_size, docs_count }
|
IndexTemplate { db_name: db_name_cloned.clone(), db_size, docs_count }
|
||||||
@ -304,7 +300,6 @@ pub fn run(opt: Opt) -> anyhow::Result<()> {
|
|||||||
|
|
||||||
let update_store_cloned = update_store.clone();
|
let update_store_cloned = update_store.clone();
|
||||||
let lmdb_path_cloned = lmdb_path.clone();
|
let lmdb_path_cloned = lmdb_path.clone();
|
||||||
let env_cloned = env.clone();
|
|
||||||
let index_cloned = index.clone();
|
let index_cloned = index.clone();
|
||||||
let updates_list_or_html_route = warp::filters::method::get()
|
let updates_list_or_html_route = warp::filters::method::get()
|
||||||
.and(warp::header("Accept"))
|
.and(warp::header("Accept"))
|
||||||
@ -335,7 +330,7 @@ pub fn run(opt: Opt) -> anyhow::Result<()> {
|
|||||||
.len() as usize;
|
.len() as usize;
|
||||||
|
|
||||||
// And the number of documents in the database.
|
// And the number of documents in the database.
|
||||||
let rtxn = env_cloned.clone().read_txn().unwrap();
|
let rtxn = index_cloned.clone().read_txn().unwrap();
|
||||||
let docs_count = index_cloned.clone().number_of_documents(&rtxn).unwrap() as usize;
|
let docs_count = index_cloned.clone().number_of_documents(&rtxn).unwrap() as usize;
|
||||||
|
|
||||||
let template = UpdatesTemplate {
|
let template = UpdatesTemplate {
|
||||||
@ -418,14 +413,13 @@ pub fn run(opt: Opt) -> anyhow::Result<()> {
|
|||||||
query: Option<String>,
|
query: Option<String>,
|
||||||
}
|
}
|
||||||
|
|
||||||
let env_cloned = env.clone();
|
|
||||||
let disable_highlighting = opt.disable_highlighting;
|
let disable_highlighting = opt.disable_highlighting;
|
||||||
let query_route = warp::filters::method::post()
|
let query_route = warp::filters::method::post()
|
||||||
.and(warp::path!("query"))
|
.and(warp::path!("query"))
|
||||||
.and(warp::body::json())
|
.and(warp::body::json())
|
||||||
.map(move |query: QueryBody| {
|
.map(move |query: QueryBody| {
|
||||||
let before_search = Instant::now();
|
let before_search = Instant::now();
|
||||||
let rtxn = env_cloned.read_txn().unwrap();
|
let rtxn = index.read_txn().unwrap();
|
||||||
|
|
||||||
let mut search = index.search(&rtxn);
|
let mut search = index.search(&rtxn);
|
||||||
if let Some(query) = query.query {
|
if let Some(query) = query.query {
|
||||||
|
@ -13,6 +13,7 @@ impl<'t, 'u, 'i> ClearDocuments<'t, 'u, 'i> {
|
|||||||
|
|
||||||
pub fn execute(self) -> anyhow::Result<usize> {
|
pub fn execute(self) -> anyhow::Result<usize> {
|
||||||
let Index {
|
let Index {
|
||||||
|
env: _env,
|
||||||
main: _main,
|
main: _main,
|
||||||
word_docids,
|
word_docids,
|
||||||
docid_word_positions,
|
docid_word_positions,
|
||||||
|
@ -69,6 +69,7 @@ impl<'t, 'u, 'i> DeleteDocuments<'t, 'u, 'i> {
|
|||||||
let id_field = fields_ids_map.id("id").expect(r#"the field "id" to be present"#);
|
let id_field = fields_ids_map.id("id").expect(r#"the field "id" to be present"#);
|
||||||
|
|
||||||
let Index {
|
let Index {
|
||||||
|
env: _env,
|
||||||
main: _main,
|
main: _main,
|
||||||
word_docids,
|
word_docids,
|
||||||
docid_word_positions,
|
docid_word_positions,
|
||||||
|
Loading…
Reference in New Issue
Block a user