mirror of
https://github.com/meilisearch/meilisearch.git
synced 2024-11-25 11:35:05 +08:00
Create a new export documents meilitool subcommand
This commit is contained in:
parent
30f3c30389
commit
feede0fad2
1
Cargo.lock
generated
1
Cargo.lock
generated
@ -3521,6 +3521,7 @@ dependencies = [
|
|||||||
"meilisearch-auth",
|
"meilisearch-auth",
|
||||||
"meilisearch-types",
|
"meilisearch-types",
|
||||||
"serde",
|
"serde",
|
||||||
|
"serde_json",
|
||||||
"time",
|
"time",
|
||||||
"uuid",
|
"uuid",
|
||||||
]
|
]
|
||||||
|
@ -16,5 +16,6 @@ file-store = { path = "../file-store" }
|
|||||||
meilisearch-auth = { path = "../meilisearch-auth" }
|
meilisearch-auth = { path = "../meilisearch-auth" }
|
||||||
meilisearch-types = { path = "../meilisearch-types" }
|
meilisearch-types = { path = "../meilisearch-types" }
|
||||||
serde = { version = "1.0.209", features = ["derive"] }
|
serde = { version = "1.0.209", features = ["derive"] }
|
||||||
|
serde_json = { version = "1.0" }
|
||||||
time = { version = "0.3.36", features = ["formatting"] }
|
time = { version = "0.3.36", features = ["formatting"] }
|
||||||
uuid = { version = "1.10.0", features = ["v4"], default-features = false }
|
uuid = { version = "1.10.0", features = ["v4"], default-features = false }
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
use std::fs::{read_dir, read_to_string, remove_file, File};
|
use std::fs::{read_dir, read_to_string, remove_file, File};
|
||||||
use std::io::BufWriter;
|
use std::io::{BufWriter, Write};
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
use anyhow::{bail, Context};
|
use anyhow::{bail, Context};
|
||||||
@ -64,6 +64,13 @@ enum Command {
|
|||||||
skip_enqueued_tasks: bool,
|
skip_enqueued_tasks: bool,
|
||||||
},
|
},
|
||||||
|
|
||||||
|
/// Exports the documents of an index from the Meilisearch database to stdout.
|
||||||
|
ExportDocuments {
|
||||||
|
/// The index name to export the documents from.
|
||||||
|
#[arg(long)]
|
||||||
|
index_name: String,
|
||||||
|
},
|
||||||
|
|
||||||
/// Attempts to upgrade from one major version to the next without a dump.
|
/// Attempts to upgrade from one major version to the next without a dump.
|
||||||
///
|
///
|
||||||
/// Make sure to run this commmand when Meilisearch is not running!
|
/// Make sure to run this commmand when Meilisearch is not running!
|
||||||
@ -89,6 +96,7 @@ fn main() -> anyhow::Result<()> {
|
|||||||
Command::ExportADump { dump_dir, skip_enqueued_tasks } => {
|
Command::ExportADump { dump_dir, skip_enqueued_tasks } => {
|
||||||
export_a_dump(db_path, dump_dir, skip_enqueued_tasks)
|
export_a_dump(db_path, dump_dir, skip_enqueued_tasks)
|
||||||
}
|
}
|
||||||
|
Command::ExportDocuments { index_name } => export_documents(db_path, index_name),
|
||||||
Command::OfflineUpgrade { target_version } => {
|
Command::OfflineUpgrade { target_version } => {
|
||||||
let target_version = parse_version(&target_version).context("While parsing `--target-version`. Make sure `--target-version` is in the format MAJOR.MINOR.PATCH")?;
|
let target_version = parse_version(&target_version).context("While parsing `--target-version`. Make sure `--target-version` is in the format MAJOR.MINOR.PATCH")?;
|
||||||
OfflineUpgrade { db_path, current_version: detected_version, target_version }.upgrade()
|
OfflineUpgrade { db_path, current_version: detected_version, target_version }.upgrade()
|
||||||
@ -605,7 +613,7 @@ fn export_a_dump(
|
|||||||
db_path: PathBuf,
|
db_path: PathBuf,
|
||||||
dump_dir: PathBuf,
|
dump_dir: PathBuf,
|
||||||
skip_enqueued_tasks: bool,
|
skip_enqueued_tasks: bool,
|
||||||
) -> Result<(), anyhow::Error> {
|
) -> anyhow::Result<()> {
|
||||||
let started_at = OffsetDateTime::now_utc();
|
let started_at = OffsetDateTime::now_utc();
|
||||||
|
|
||||||
// 1. Extracts the instance UID from disk
|
// 1. Extracts the instance UID from disk
|
||||||
@ -750,3 +758,40 @@ fn export_a_dump(
|
|||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn export_documents(db_path: PathBuf, index_name: String) -> anyhow::Result<()> {
|
||||||
|
let index_scheduler_path = db_path.join("tasks");
|
||||||
|
let env = unsafe { EnvOpenOptions::new().max_dbs(100).open(&index_scheduler_path) }
|
||||||
|
.with_context(|| format!("While trying to open {:?}", index_scheduler_path.display()))?;
|
||||||
|
|
||||||
|
let rtxn = env.read_txn()?;
|
||||||
|
let index_mapping: Database<Str, UuidCodec> =
|
||||||
|
try_opening_database(&env, &rtxn, "index-mapping")?;
|
||||||
|
|
||||||
|
for result in index_mapping.iter(&rtxn)? {
|
||||||
|
let (uid, uuid) = result?;
|
||||||
|
if uid == index_name {
|
||||||
|
let index_path = db_path.join("indexes").join(uuid.to_string());
|
||||||
|
let index = Index::new(EnvOpenOptions::new(), &index_path).with_context(|| {
|
||||||
|
format!("While trying to open the index at path {:?}", index_path.display())
|
||||||
|
})?;
|
||||||
|
|
||||||
|
let rtxn = index.read_txn()?;
|
||||||
|
let fields_ids_map = index.fields_ids_map(&rtxn)?;
|
||||||
|
let all_fields: Vec<_> = fields_ids_map.iter().map(|(id, _)| id).collect();
|
||||||
|
|
||||||
|
let mut stdout = BufWriter::new(std::io::stdout());
|
||||||
|
for ret in index.all_documents(&rtxn)? {
|
||||||
|
let (_id, doc) = ret?;
|
||||||
|
let document = obkv_to_json(&all_fields, &fields_ids_map, doc)?;
|
||||||
|
serde_json::to_writer(&mut stdout, &document)?;
|
||||||
|
}
|
||||||
|
|
||||||
|
stdout.flush()?;
|
||||||
|
} else {
|
||||||
|
eprintln!("Found index {uid} but it's not the right index...");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user