From 64ef5869d72c6b962c20b55740e2260d042520d9 Mon Sep 17 00:00:00 2001 From: many Date: Thu, 18 Nov 2021 16:56:05 +0100 Subject: [PATCH 1/2] Update tokenizer v0.2.6 --- http-ui/Cargo.toml | 2 +- milli/Cargo.toml | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/http-ui/Cargo.toml b/http-ui/Cargo.toml index 04e1c708a..d807c4923 100644 --- a/http-ui/Cargo.toml +++ b/http-ui/Cargo.toml @@ -10,7 +10,7 @@ anyhow = "1.0.38" byte-unit = { version = "4.0.9", default-features = false, features = ["std"] } crossbeam-channel = "0.5.0" heed = { git = "https://github.com/Kerollmops/heed", tag = "v0.12.1" } -meilisearch-tokenizer = { git = "https://github.com/meilisearch/tokenizer.git", tag = "v0.2.5" } +meilisearch-tokenizer = { git = "https://github.com/meilisearch/tokenizer.git", tag = "v0.2.6" } memmap2 = "0.5.0" milli = { path = "../milli" } once_cell = "1.5.2" diff --git a/milli/Cargo.toml b/milli/Cargo.toml index 90bd1f926..790b52647 100644 --- a/milli/Cargo.toml +++ b/milli/Cargo.toml @@ -22,7 +22,7 @@ heed = { git = "https://github.com/Kerollmops/heed", tag = "v0.12.1", default-fe human_format = "1.0.3" levenshtein_automata = { version = "0.2.0", features = ["fst_automaton"] } linked-hash-map = "0.5.4" -meilisearch-tokenizer = { git = "https://github.com/meilisearch/tokenizer.git", tag = "v0.2.5" } +meilisearch-tokenizer = { git = "https://github.com/meilisearch/tokenizer.git", tag = "v0.2.6" } memmap2 = "0.5.0" obkv = "0.2.0" once_cell = "1.5.2" From 35f949963821a1c6ae43cb1a660b087d5ea36c0c Mon Sep 17 00:00:00 2001 From: many Date: Thu, 18 Nov 2021 16:57:12 +0100 Subject: [PATCH 2/2] Export tokenizer from milli --- milli/src/lib.rs | 1 + 1 file changed, 1 insertion(+) diff --git a/milli/src/lib.rs b/milli/src/lib.rs index 044d74ec1..9e7bb8966 100644 --- a/milli/src/lib.rs +++ b/milli/src/lib.rs @@ -20,6 +20,7 @@ use std::hash::BuildHasherDefault; pub use filter_parser::{Condition, FilterCondition}; use fxhash::{FxHasher32, FxHasher64}; pub use grenad::CompressionType; +pub use meilisearch_tokenizer as tokenizer; use serde_json::{Map, Value}; pub use self::asc_desc::{AscDesc, AscDescError, Member, SortError};