mirror of
https://github.com/meilisearch/meilisearch.git
synced 2024-11-23 02:27:40 +08:00
3170: Re-enable importing from dumps v1 r=irevoire a=dureuill # Pull Request ## Related issue Fixes #2985 ## What does this PR do? ### User standpoint - Allows importing dumps version 1 (exported with Meilisearch <=v0.20) to modern-day Meilisearch - Tasks of type "Customs" are skipped, with a warning - Tasks of status "enqueued" are skipped, with a warning - The "WordsPosition" ranking rule is skipped when encountered in the ranking rules, with a warning. After an import from a v1 dump, it is recommended that a user checks each index and its settings. ### Implementation standpoint - Add a dump v1 reader based on the one by `@irevoire` - Add a v1_to_v2 compatibility layer based on the v2_to_v3 one - as v2 requires UUIDs, the v1 indexes are mapped to UUIDs built from their position in the metada file: the first index is given UUID all zeroes, the second one UUID `00000000-0000-0000-0000-000000000001`, and so on... This should have no bearing on the final indexes because v6 is not using UUIDs, but this allows us to correctly identify which tasks belong to which index. - Modify the v2_to_v3 compatibility layer to account for the fact that the reader can actually be a v1_to_v2 compat layer - Make some base dump types Clone - impl Display for v2::settings::Criterion ## PR checklist Please check if your PR fulfills the following requirements: - [x] Does this PR fix an existing issue, or have you listed the changes applied in the PR description (and why they are needed)? - [x] Have you read the contributing guidelines? - [x] Have you made sure that the title is accurate and descriptive of the changes? Thank you so much for contributing to Meilisearch! 3180: Bump mislav/bump-homebrew-formula-action from 1 to 2 r=curquiza a=dependabot[bot] Bumps [mislav/bump-homebrew-formula-action](https://github.com/mislav/bump-homebrew-formula-action) from 1 to 2. <details> <summary>Release notes</summary> <p><em>Sourced from <a href="https://github.com/mislav/bump-homebrew-formula-action/releases">mislav/bump-homebrew-formula-action's releases</a>.</em></p> <blockquote> <h2>bump-homebrew-formula 2.0</h2> <h2>What's Changed</h2> <ul> <li>Use Node 16 by <a href="https://github.com/chenrui333"><code>`@chenrui333</code></a>` in <a href="https://github-redirect.dependabot.com/mislav/bump-homebrew-formula-action/pull/36">mislav/bump-homebrew-formula-action#36</a></li> <li>Bump minimist from 1.2.5 to 1.2.6 by <a href="https://github.com/dependabot"><code>`@dependabot</code></a>` in <a href="https://github-redirect.dependabot.com/mislav/bump-homebrew-formula-action/pull/33">mislav/bump-homebrew-formula-action#33</a></li> <li>Bump node-fetch from 2.6.6 to 2.6.7 by <a href="https://github.com/dependabot"><code>`@dependabot</code></a>` in <a href="https://github-redirect.dependabot.com/mislav/bump-homebrew-formula-action/pull/34">mislav/bump-homebrew-formula-action#34</a></li> </ul> <h2>New Contributors</h2> <ul> <li><a href="https://github.com/chenrui333"><code>`@chenrui333</code></a>` made their first contribution in <a href="https://github-redirect.dependabot.com/mislav/bump-homebrew-formula-action/pull/36">mislav/bump-homebrew-formula-action#36</a></li> </ul> <h2>bump-homebrew-formula 1.16</h2> <ul> <li>Replaces broken v1.15 tag, thanks <a href="https://github.com/hendrikmaus"><code>`@hendrikmaus</code></a>` <a href="https://github-redirect.dependabot.com/mislav/bump-homebrew-formula-action/issues/32">mislav/bump-homebrew-formula-action#32</a></li> <li>Add <code>push-to</code> option, thanks <a href="https://github.com/codefromthecrypt"><code>`@codefromthecrypt</code></a>` <a href="https://github-redirect.dependabot.com/mislav/bump-homebrew-formula-action/pull/30">mislav/bump-homebrew-formula-action#30</a></li> <li>Fix syntax error, thanks <a href="https://github.com/hendrikmaus"><code>`@hendrikmaus</code></a>` <a href="https://github.com/wata727"><code>`@wata727</code></a>` <a href="https://github-redirect.dependabot.com/mislav/bump-homebrew-formula-action/pull/27">mislav/bump-homebrew-formula-action#27</a></li> <li>Ensure repeated placeholders in <code>commit-message</code> are expanded, thanks <a href="https://github.com/hendrikmaus"><code>`@hendrikmaus</code></a>` <a href="https://github-redirect.dependabot.com/mislav/bump-homebrew-formula-action/pull/29">mislav/bump-homebrew-formula-action#29</a></li> </ul> <h2>bump-homebrew-formula 1.14</h2> <ul> <li>Ignore HTTP 409 error when fast-forwading the main branch of <code>homebrew-tap</code> fork</li> </ul> <h2>bump-homebrew-formula 1.13</h2> <ul> <li>Add <code>create-pullrequest</code> input to control whether or not a PR is submitted to <code>homebrew-tap</code></li> <li>Add <code>download-sha256</code> input to define the SHA256 checksum of the archive at <code>download-url</code></li> <li>Fix creating a new branch in the forked repo failing with HTTP 404</li> </ul> <h2>bump-homebrew-formula 1.12</h2> <ul> <li>Fix Actions CJS loader halting on <code>foo?.bar</code> JS syntax</li> </ul> <h2>bump-homebrew-formula 1.11</h2> <ul> <li>New optional <code>formula-path</code> input accepts the filename of the formula file to edit (default <code>Formula/<formula-name>.rb</code>).</li> <li>Remove <code>revision N</code> lines when bumping Homebrew formulae.</li> </ul> <h2>bump-homebrew-formula 1.10</h2> <ul> <li>The new optional <code>tag-name</code> input allows this action to be <a href="https://docs.github.com/en/actions/managing-workflow-runs/manually-running-a-workflow">manually triggered via <code>workflow_dispatch</code></a> instead of on git push to a tag.</li> </ul> <h2>bump-homebrew-formula 1.9</h2> <ul> <li>Fix following multiple HTTP redirects while calculating checksum for <code>download-url</code></li> </ul> <h2>bump-homebrew-formula 1.8</h2> <ul> <li>Enable JavaScript source maps for better failure debugging</li> </ul> <h2>bump-homebrew-formula 1.7</h2> <ul> <li> <p>Allow <code>download-url</code> as input parameter</p> </li> <li> <p>Add support for git-based <code>download-url</code></p> </li> </ul> <h2>bump-homebrew-formula 1.6</h2> <ul> <li>Control the git commit message template being used for updating the formula file via the <code>commit-message</code> action input</li> </ul> <h2>bump-homebrew-formula 1.5</h2> <ul> <li>Support detection version from <code>https://github.com/OWNER/REPO/releases/download/TAG/FILE</code> download URLs</li> </ul> <!-- raw HTML omitted --> </blockquote> <p>... (truncated)</p> </details> <details> <summary>Commits</summary> <ul> <li><a href="fcd7e28e54
"><code>fcd7e28</code></a> lib</li> <li><a href="33989a8502
"><code>33989a8</code></a> Merge branch 'main' into v2</li> <li><a href="5983bb6c59
"><code>5983bb6</code></a> Improve extracting complex tag names from URLs</li> <li><a href="9750a1166b
"><code>9750a11</code></a> lib</li> <li><a href="64410e9c96
"><code>64410e9</code></a> v2</li> <li><a href="677d7482a3
"><code>677d748</code></a> Merge pull request <a href="https://github-redirect.dependabot.com/mislav/bump-homebrew-formula-action/issues/36">#36</a> from chenrui333/node-16</li> <li><a href="f364e76079
"><code>f364e76</code></a> also update some build dependencies</li> <li><a href="c08fd9bee5
"><code>c08fd9b</code></a> deps: update to use nodev16</li> <li><a href="280f532e9a
"><code>280f532</code></a> Merge pull request <a href="https://github-redirect.dependabot.com/mislav/bump-homebrew-formula-action/issues/34">#34</a> from mislav/dependabot/npm_and_yarn/node-fetch-2.6.7</li> <li><a href="5d94a66af3
"><code>5d94a66</code></a> Bump node-fetch from 2.6.6 to 2.6.7</li> <li>Additional commits viewable in <a href="https://github.com/mislav/bump-homebrew-formula-action/compare/v1...v2">compare view</a></li> </ul> </details> <br /> [![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=mislav/bump-homebrew-formula-action&package-manager=github_actions&previous-version=1&new-version=2)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) You can trigger a rebase of this PR by commenting ``@dependabot` rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) --- <details> <summary>Dependabot commands and options</summary> <br /> You can trigger Dependabot actions by commenting on this PR: - ``@dependabot` rebase` will rebase this PR - ``@dependabot` recreate` will recreate this PR, overwriting any edits that have been made to it - ``@dependabot` merge` will merge this PR after your CI passes on it - ``@dependabot` squash and merge` will squash and merge this PR after your CI passes on it - ``@dependabot` cancel merge` will cancel a previously requested merge and block automerging - ``@dependabot` reopen` will reopen this PR if it is closed - ``@dependabot` close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - ``@dependabot` ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - ``@dependabot` ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - ``@dependabot` ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself) </details> 3181: Bump Swatinem/rust-cache from 2.0.0 to 2.2.0 r=curquiza a=dependabot[bot] Bumps [Swatinem/rust-cache](https://github.com/Swatinem/rust-cache) from 2.0.0 to 2.2.0. <details> <summary>Release notes</summary> <p><em>Sourced from <a href="https://github.com/Swatinem/rust-cache/releases">Swatinem/rust-cache's releases</a>.</em></p> <blockquote> <h2>v2.2.0</h2> <ul> <li>Add new <code>save-if</code> option to always restore, but only conditionally save the cache.</li> </ul> <h2>v2.1.0</h2> <ul> <li>Only hash <code>Cargo.{lock,toml}</code> files in the configured workspace directories.</li> </ul> <h2>v2.0.2</h2> <ul> <li>Avoid calling cargo metadata on pre-cleanup.</li> <li>Added <code>prefix-key</code>, <code>cache-directories</code> and <code>cache-targets</code> options.</li> </ul> <h2>v2.0.1</h2> <ul> <li>Primarily just updating dependencies to fix GitHub deprecation notices.</li> </ul> </blockquote> </details> <details> <summary>Changelog</summary> <p><em>Sourced from <a href="https://github.com/Swatinem/rust-cache/blob/master/CHANGELOG.md">Swatinem/rust-cache's changelog</a>.</em></p> <blockquote> <h2>2.2.0</h2> <ul> <li>Add new <code>save-if</code> option to always restore, but only conditionally save the cache.</li> </ul> <h2>2.1.0</h2> <ul> <li>Only hash <code>Cargo.{lock,toml}</code> files in the configured workspace directories.</li> </ul> <h2>2.0.2</h2> <ul> <li>Avoid calling <code>cargo metadata</code> on pre-cleanup.</li> <li>Added <code>prefix-key</code>, <code>cache-directories</code> and <code>cache-targets</code> options.</li> </ul> <h2>2.0.1</h2> <ul> <li>Primarily just updating dependencies to fix GitHub deprecation notices.</li> </ul> </blockquote> </details> <details> <summary>Commits</summary> <ul> <li><a href="359a70e43a
"><code>359a70e</code></a> 2.2.0</li> <li><a href="ecee04e7b3
"><code>ecee04e</code></a> feat: add save-if option, closes <a href="https://github-redirect.dependabot.com/Swatinem/rust-cache/issues/66">#66</a> (<a href="https://github-redirect.dependabot.com/Swatinem/rust-cache/issues/91">#91</a>)</li> <li><a href="b894d59a8d
"><code>b894d59</code></a> 2.1.0</li> <li><a href="e78327dd9e
"><code>e78327d</code></a> small code style improvements, README and CHANGELOG updates</li> <li><a href="ccdddcc049
"><code>ccdddcc</code></a> only hash Cargo.toml/Cargo.lock that belong to a configured workspace (<a href="https://github-redirect.dependabot.com/Swatinem/rust-cache/issues/90">#90</a>)</li> <li><a href="b5ec9edd91
"><code>b5ec9ed</code></a> 2.0.2</li> <li><a href="3f2513fdf4
"><code>3f2513f</code></a> avoid calling cargo metadata on pre-cleanup</li> <li><a href="19c46583c5
"><code>19c4658</code></a> update dependencies</li> <li><a href="b8e72aae83
"><code>b8e72aa</code></a> Added <code>prefix-key</code> <code>cache-directories</code> and <code>cache-targets</code> options (<a href="https://github-redirect.dependabot.com/Swatinem/rust-cache/issues/85">#85</a>)</li> <li><a href="22c9328bcb
"><code>22c9328</code></a> 2.0.1</li> <li>Additional commits viewable in <a href="https://github.com/Swatinem/rust-cache/compare/v2.0.0...v2.2.0">compare view</a></li> </ul> </details> <br /> [![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=Swatinem/rust-cache&package-manager=github_actions&previous-version=2.0.0&new-version=2.2.0)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) You can trigger a rebase of this PR by commenting ``@dependabot` rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) --- <details> <summary>Dependabot commands and options</summary> <br /> You can trigger Dependabot actions by commenting on this PR: - ``@dependabot` rebase` will rebase this PR - ``@dependabot` recreate` will recreate this PR, overwriting any edits that have been made to it - ``@dependabot` merge` will merge this PR after your CI passes on it - ``@dependabot` squash and merge` will squash and merge this PR after your CI passes on it - ``@dependabot` cancel merge` will cancel a previously requested merge and block automerging - ``@dependabot` reopen` will reopen this PR if it is closed - ``@dependabot` close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - ``@dependabot` ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - ``@dependabot` ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - ``@dependabot` ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself) </details> 3183: Use ubuntu-latest when not impacting r=Kerollmops a=curquiza Minor changes - Use `ubuntu-latest` for CI where there is no compilation - rename one of the workflow (obsolete name) Co-authored-by: Louis Dureuil <louis@meilisearch.com> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: curquiza <clementine@meilisearch.com>
This commit is contained in:
commit
40339919ad
2
.github/workflows/coverage.yml
vendored
2
.github/workflows/coverage.yml
vendored
@ -6,7 +6,7 @@ name: Execute code coverage
|
|||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
nightly-coverage:
|
nightly-coverage:
|
||||||
runs-on: ubuntu-18.04
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v3
|
||||||
- uses: actions-rs/toolchain@v1
|
- uses: actions-rs/toolchain@v1
|
||||||
|
6
.github/workflows/publish-deb-brew-pkg.yml
vendored
6
.github/workflows/publish-deb-brew-pkg.yml
vendored
@ -1,4 +1,4 @@
|
|||||||
name: Publish deb pkg to GitHub release & APT repository & Homebrew
|
name: Publish to APT repository & Homebrew
|
||||||
|
|
||||||
on:
|
on:
|
||||||
release:
|
release:
|
||||||
@ -38,11 +38,11 @@ jobs:
|
|||||||
|
|
||||||
homebrew:
|
homebrew:
|
||||||
name: Bump Homebrew formula
|
name: Bump Homebrew formula
|
||||||
runs-on: ubuntu-18.04
|
runs-on: ubuntu-latest
|
||||||
needs: check-version
|
needs: check-version
|
||||||
steps:
|
steps:
|
||||||
- name: Create PR to Homebrew
|
- name: Create PR to Homebrew
|
||||||
uses: mislav/bump-homebrew-formula-action@v1
|
uses: mislav/bump-homebrew-formula-action@v2
|
||||||
with:
|
with:
|
||||||
formula-name: meilisearch
|
formula-name: meilisearch
|
||||||
env:
|
env:
|
||||||
|
8
.github/workflows/rust.yml
vendored
8
.github/workflows/rust.yml
vendored
@ -25,7 +25,7 @@ jobs:
|
|||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v3
|
||||||
- name: Cache dependencies
|
- name: Cache dependencies
|
||||||
uses: Swatinem/rust-cache@v2.0.0
|
uses: Swatinem/rust-cache@v2.2.0
|
||||||
- name: Run cargo check without any default features
|
- name: Run cargo check without any default features
|
||||||
uses: actions-rs/cargo@v1
|
uses: actions-rs/cargo@v1
|
||||||
with:
|
with:
|
||||||
@ -49,7 +49,7 @@ jobs:
|
|||||||
toolchain: stable
|
toolchain: stable
|
||||||
override: true
|
override: true
|
||||||
- name: Cache dependencies
|
- name: Cache dependencies
|
||||||
uses: Swatinem/rust-cache@v2.0.0
|
uses: Swatinem/rust-cache@v2.2.0
|
||||||
- name: Run tests in debug
|
- name: Run tests in debug
|
||||||
uses: actions-rs/cargo@v1
|
uses: actions-rs/cargo@v1
|
||||||
with:
|
with:
|
||||||
@ -68,7 +68,7 @@ jobs:
|
|||||||
override: true
|
override: true
|
||||||
components: clippy
|
components: clippy
|
||||||
- name: Cache dependencies
|
- name: Cache dependencies
|
||||||
uses: Swatinem/rust-cache@v2.0.0
|
uses: Swatinem/rust-cache@v2.2.0
|
||||||
- name: Run cargo clippy
|
- name: Run cargo clippy
|
||||||
uses: actions-rs/cargo@v1
|
uses: actions-rs/cargo@v1
|
||||||
with:
|
with:
|
||||||
@ -87,6 +87,6 @@ jobs:
|
|||||||
override: true
|
override: true
|
||||||
components: rustfmt
|
components: rustfmt
|
||||||
- name: Cache dependencies
|
- name: Cache dependencies
|
||||||
uses: Swatinem/rust-cache@v2.0.0
|
uses: Swatinem/rust-cache@v2.2.0
|
||||||
- name: Run cargo fmt
|
- name: Run cargo fmt
|
||||||
run: cargo fmt --all -- --check
|
run: cargo fmt --all -- --check
|
||||||
|
@ -3,8 +3,6 @@ use thiserror::Error;
|
|||||||
|
|
||||||
#[derive(Debug, Error)]
|
#[derive(Debug, Error)]
|
||||||
pub enum Error {
|
pub enum Error {
|
||||||
#[error("The version 1 of the dumps is not supported anymore. You can re-export your dump from a version between 0.21 and 0.24, or start fresh from a version 0.25 onwards.")]
|
|
||||||
DumpV1Unsupported,
|
|
||||||
#[error("Bad index name.")]
|
#[error("Bad index name.")]
|
||||||
BadIndexName,
|
BadIndexName,
|
||||||
#[error("Malformed task.")]
|
#[error("Malformed task.")]
|
||||||
@ -28,7 +26,6 @@ impl ErrorCode for Error {
|
|||||||
Error::Uuid(_) => Code::Internal,
|
Error::Uuid(_) => Code::Internal,
|
||||||
|
|
||||||
// all these errors should never be raised when creating a dump, thus no error code should be associated.
|
// all these errors should never be raised when creating a dump, thus no error code should be associated.
|
||||||
Error::DumpV1Unsupported => Code::Internal,
|
|
||||||
Error::BadIndexName => Code::Internal,
|
Error::BadIndexName => Code::Internal,
|
||||||
Error::MalformedTask => Code::Internal,
|
Error::MalformedTask => Code::Internal,
|
||||||
}
|
}
|
||||||
|
@ -23,7 +23,7 @@ const CURRENT_DUMP_VERSION: Version = Version::V6;
|
|||||||
|
|
||||||
type Result<T> = std::result::Result<T, Error>;
|
type Result<T> = std::result::Result<T, Error>;
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Serialize, Deserialize)]
|
#[derive(Debug, Clone, PartialEq, Eq, Serialize, Deserialize)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
pub struct Metadata {
|
pub struct Metadata {
|
||||||
pub dump_version: Version,
|
pub dump_version: Version,
|
||||||
@ -32,7 +32,7 @@ pub struct Metadata {
|
|||||||
pub dump_date: OffsetDateTime,
|
pub dump_date: OffsetDateTime,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Serialize, Deserialize)]
|
#[derive(Debug, Clone, PartialEq, Eq, Serialize, Deserialize)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
pub struct IndexMetadata {
|
pub struct IndexMetadata {
|
||||||
pub uid: String,
|
pub uid: String,
|
||||||
@ -43,7 +43,7 @@ pub struct IndexMetadata {
|
|||||||
pub updated_at: OffsetDateTime,
|
pub updated_at: OffsetDateTime,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Deserialize, Serialize)]
|
#[derive(Debug, Clone, Copy, PartialEq, Eq, Deserialize, Serialize)]
|
||||||
pub enum Version {
|
pub enum Version {
|
||||||
V1,
|
V1,
|
||||||
V2,
|
V2,
|
||||||
|
@ -1,3 +1,4 @@
|
|||||||
|
pub mod v1_to_v2;
|
||||||
pub mod v2_to_v3;
|
pub mod v2_to_v3;
|
||||||
pub mod v3_to_v4;
|
pub mod v3_to_v4;
|
||||||
pub mod v4_to_v5;
|
pub mod v4_to_v5;
|
||||||
|
@ -0,0 +1,23 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/compat/v1_to_v2.rs
|
||||||
|
expression: spells.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"filterableAttributes": [],
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"exactness"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"distinctAttribute": null
|
||||||
|
}
|
@ -0,0 +1,37 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/compat/v1_to_v2.rs
|
||||||
|
expression: products.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"filterableAttributes": [],
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"exactness"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {
|
||||||
|
"android": [
|
||||||
|
"phone",
|
||||||
|
"smartphone"
|
||||||
|
],
|
||||||
|
"iphone": [
|
||||||
|
"phone",
|
||||||
|
"smartphone"
|
||||||
|
],
|
||||||
|
"phone": [
|
||||||
|
"android",
|
||||||
|
"iphone",
|
||||||
|
"smartphone"
|
||||||
|
]
|
||||||
|
},
|
||||||
|
"distinctAttribute": null
|
||||||
|
}
|
@ -0,0 +1,27 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/compat/v1_to_v2.rs
|
||||||
|
expression: movies.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"filterableAttributes": [
|
||||||
|
"genres",
|
||||||
|
"id"
|
||||||
|
],
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"exactness",
|
||||||
|
"asc(release_date)"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"distinctAttribute": null
|
||||||
|
}
|
@ -0,0 +1,27 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/compat/v1_to_v2.rs
|
||||||
|
expression: movies.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"filterableAttributes": [
|
||||||
|
"genres",
|
||||||
|
"id"
|
||||||
|
],
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"exactness",
|
||||||
|
"asc(release_date)"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"distinctAttribute": null
|
||||||
|
}
|
@ -0,0 +1,23 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/compat/v1_to_v2.rs
|
||||||
|
expression: spells.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"filterableAttributes": [],
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"exactness"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"distinctAttribute": null
|
||||||
|
}
|
414
dump/src/reader/compat/v1_to_v2.rs
Normal file
414
dump/src/reader/compat/v1_to_v2.rs
Normal file
@ -0,0 +1,414 @@
|
|||||||
|
use std::{collections::BTreeSet, str::FromStr};
|
||||||
|
|
||||||
|
use crate::reader::{v1, v2, Document};
|
||||||
|
|
||||||
|
use super::v2_to_v3::CompatV2ToV3;
|
||||||
|
use crate::Result;
|
||||||
|
|
||||||
|
pub struct CompatV1ToV2 {
|
||||||
|
pub from: v1::V1Reader,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl CompatV1ToV2 {
|
||||||
|
pub fn new(v1: v1::V1Reader) -> Self {
|
||||||
|
Self { from: v1 }
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn to_v3(self) -> CompatV2ToV3 {
|
||||||
|
CompatV2ToV3::Compat(self)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn version(&self) -> crate::Version {
|
||||||
|
self.from.version()
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn date(&self) -> Option<time::OffsetDateTime> {
|
||||||
|
self.from.date()
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn index_uuid(&self) -> Vec<v2::meta::IndexUuid> {
|
||||||
|
self.from
|
||||||
|
.index_uuid()
|
||||||
|
.into_iter()
|
||||||
|
.enumerate()
|
||||||
|
// we use the index of the index 😬 as UUID for the index, so that we can link the v2::Task to their index
|
||||||
|
.map(|(index, index_uuid)| v2::meta::IndexUuid {
|
||||||
|
uid: index_uuid.uid,
|
||||||
|
uuid: uuid::Uuid::from_u128(index as u128),
|
||||||
|
})
|
||||||
|
.collect()
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn indexes(&self) -> Result<impl Iterator<Item = Result<CompatIndexV1ToV2>> + '_> {
|
||||||
|
Ok(self.from.indexes()?.map(|index_reader| Ok(CompatIndexV1ToV2 { from: index_reader? })))
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn tasks(
|
||||||
|
&mut self,
|
||||||
|
) -> Box<dyn Iterator<Item = Result<(v2::Task, Option<v2::UpdateFile>)>> + '_> {
|
||||||
|
// Convert an error here to an iterator yielding the error
|
||||||
|
let indexes = match self.from.indexes() {
|
||||||
|
Ok(indexes) => indexes,
|
||||||
|
Err(err) => return Box::new(std::iter::once(Err(err))),
|
||||||
|
};
|
||||||
|
let it = indexes.enumerate().flat_map(
|
||||||
|
move |(index, index_reader)| -> Box<dyn Iterator<Item = _>> {
|
||||||
|
let index_reader = match index_reader {
|
||||||
|
Ok(index_reader) => index_reader,
|
||||||
|
Err(err) => return Box::new(std::iter::once(Err(err))),
|
||||||
|
};
|
||||||
|
Box::new(
|
||||||
|
index_reader
|
||||||
|
.tasks()
|
||||||
|
// Filter out the UpdateStatus::Customs variant that is not supported in v2
|
||||||
|
// and enqueued tasks, that don't contain the necessary update file in v1
|
||||||
|
.filter_map(move |task| -> Option<_> {
|
||||||
|
let task = match task {
|
||||||
|
Ok(task) => task,
|
||||||
|
Err(err) => return Some(Err(err)),
|
||||||
|
};
|
||||||
|
Some(Ok((
|
||||||
|
v2::Task {
|
||||||
|
uuid: uuid::Uuid::from_u128(index as u128),
|
||||||
|
update: Option::from(task)?,
|
||||||
|
},
|
||||||
|
None,
|
||||||
|
)))
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
},
|
||||||
|
);
|
||||||
|
Box::new(it)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct CompatIndexV1ToV2 {
|
||||||
|
pub from: v1::V1IndexReader,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl CompatIndexV1ToV2 {
|
||||||
|
pub fn metadata(&self) -> &crate::IndexMetadata {
|
||||||
|
self.from.metadata()
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn documents(&mut self) -> Result<Box<dyn Iterator<Item = Result<Document>> + '_>> {
|
||||||
|
self.from.documents().map(|it| Box::new(it) as Box<dyn Iterator<Item = _>>)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn settings(&mut self) -> Result<v2::settings::Settings<v2::settings::Checked>> {
|
||||||
|
Ok(v2::settings::Settings::<v2::settings::Unchecked>::from(self.from.settings()?).check())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<v1::settings::Settings> for v2::Settings<v2::Unchecked> {
|
||||||
|
fn from(source: v1::settings::Settings) -> Self {
|
||||||
|
let displayed_attributes = source
|
||||||
|
.displayed_attributes
|
||||||
|
.map(|opt| opt.map(|displayed_attributes| displayed_attributes.into_iter().collect()));
|
||||||
|
let attributes_for_faceting = source.attributes_for_faceting.map(|opt| {
|
||||||
|
opt.map(|attributes_for_faceting| attributes_for_faceting.into_iter().collect())
|
||||||
|
});
|
||||||
|
let ranking_rules = source.ranking_rules.map(|opt| {
|
||||||
|
opt.map(|ranking_rules| {
|
||||||
|
ranking_rules
|
||||||
|
.into_iter()
|
||||||
|
.filter_map(|ranking_rule| {
|
||||||
|
match v1::settings::RankingRule::from_str(&ranking_rule) {
|
||||||
|
Ok(ranking_rule) => {
|
||||||
|
let criterion: Option<v2::settings::Criterion> =
|
||||||
|
ranking_rule.into();
|
||||||
|
criterion.as_ref().map(ToString::to_string)
|
||||||
|
}
|
||||||
|
Err(()) => Some(ranking_rule),
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.collect()
|
||||||
|
})
|
||||||
|
});
|
||||||
|
|
||||||
|
Self {
|
||||||
|
displayed_attributes,
|
||||||
|
searchable_attributes: source.searchable_attributes,
|
||||||
|
filterable_attributes: attributes_for_faceting,
|
||||||
|
ranking_rules,
|
||||||
|
stop_words: source.stop_words,
|
||||||
|
synonyms: source.synonyms,
|
||||||
|
distinct_attribute: source.distinct_attribute,
|
||||||
|
_kind: std::marker::PhantomData,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<v1::update::UpdateStatus> for Option<v2::updates::UpdateStatus> {
|
||||||
|
fn from(source: v1::update::UpdateStatus) -> Self {
|
||||||
|
use v1::update::UpdateStatus as UpdateStatusV1;
|
||||||
|
use v2::updates::UpdateStatus as UpdateStatusV2;
|
||||||
|
Some(match source {
|
||||||
|
UpdateStatusV1::Enqueued { content } => {
|
||||||
|
log::warn!(
|
||||||
|
"Cannot import task {} (importing enqueued tasks from v1 dumps is unsupported)",
|
||||||
|
content.update_id
|
||||||
|
);
|
||||||
|
log::warn!("Task will be skipped in the queue of imported tasks.");
|
||||||
|
|
||||||
|
return None;
|
||||||
|
}
|
||||||
|
UpdateStatusV1::Failed { content } => UpdateStatusV2::Failed(v2::updates::Failed {
|
||||||
|
from: v2::updates::Processing {
|
||||||
|
from: v2::updates::Enqueued {
|
||||||
|
update_id: content.update_id,
|
||||||
|
meta: Option::from(content.update_type)?,
|
||||||
|
enqueued_at: content.enqueued_at,
|
||||||
|
content: None,
|
||||||
|
},
|
||||||
|
started_processing_at: content.processed_at
|
||||||
|
- std::time::Duration::from_secs_f64(content.duration),
|
||||||
|
},
|
||||||
|
error: v2::ResponseError {
|
||||||
|
// error code is ignored by serialization, and so always default in deserialized v2 dumps
|
||||||
|
// that's a good thing, because we don't have them in v1 dump 😅
|
||||||
|
code: http::StatusCode::default(),
|
||||||
|
message: content.error.unwrap_or_default(),
|
||||||
|
// error codes are unchanged between v1 and v2
|
||||||
|
error_code: content.error_code.unwrap_or_default(),
|
||||||
|
// error types are unchanged between v1 and v2
|
||||||
|
error_type: content.error_type.unwrap_or_default(),
|
||||||
|
// error links are unchanged between v1 and v2
|
||||||
|
error_link: content.error_link.unwrap_or_default(),
|
||||||
|
},
|
||||||
|
failed_at: content.processed_at,
|
||||||
|
}),
|
||||||
|
UpdateStatusV1::Processed { content } => {
|
||||||
|
UpdateStatusV2::Processed(v2::updates::Processed {
|
||||||
|
success: match &content.update_type {
|
||||||
|
v1::update::UpdateType::ClearAll => {
|
||||||
|
v2::updates::UpdateResult::DocumentDeletion { deleted: u64::MAX }
|
||||||
|
}
|
||||||
|
v1::update::UpdateType::Customs => v2::updates::UpdateResult::Other,
|
||||||
|
v1::update::UpdateType::DocumentsAddition { number } => {
|
||||||
|
v2::updates::UpdateResult::DocumentsAddition(
|
||||||
|
v2::updates::DocumentAdditionResult { nb_documents: *number },
|
||||||
|
)
|
||||||
|
}
|
||||||
|
v1::update::UpdateType::DocumentsPartial { number } => {
|
||||||
|
v2::updates::UpdateResult::DocumentsAddition(
|
||||||
|
v2::updates::DocumentAdditionResult { nb_documents: *number },
|
||||||
|
)
|
||||||
|
}
|
||||||
|
v1::update::UpdateType::DocumentsDeletion { number } => {
|
||||||
|
v2::updates::UpdateResult::DocumentDeletion { deleted: *number as u64 }
|
||||||
|
}
|
||||||
|
v1::update::UpdateType::Settings { .. } => v2::updates::UpdateResult::Other,
|
||||||
|
},
|
||||||
|
processed_at: content.processed_at,
|
||||||
|
from: v2::updates::Processing {
|
||||||
|
from: v2::updates::Enqueued {
|
||||||
|
update_id: content.update_id,
|
||||||
|
meta: Option::from(content.update_type)?,
|
||||||
|
enqueued_at: content.enqueued_at,
|
||||||
|
content: None,
|
||||||
|
},
|
||||||
|
started_processing_at: content.processed_at
|
||||||
|
- std::time::Duration::from_secs_f64(content.duration),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<v1::update::UpdateType> for Option<v2::updates::UpdateMeta> {
|
||||||
|
fn from(source: v1::update::UpdateType) -> Self {
|
||||||
|
Some(match source {
|
||||||
|
v1::update::UpdateType::ClearAll => v2::updates::UpdateMeta::ClearDocuments,
|
||||||
|
v1::update::UpdateType::Customs => {
|
||||||
|
log::warn!("Ignoring task with type 'Customs' that is no longer supported");
|
||||||
|
return None;
|
||||||
|
}
|
||||||
|
v1::update::UpdateType::DocumentsAddition { .. } => {
|
||||||
|
v2::updates::UpdateMeta::DocumentsAddition {
|
||||||
|
method: v2::updates::IndexDocumentsMethod::ReplaceDocuments,
|
||||||
|
format: v2::updates::UpdateFormat::Json,
|
||||||
|
primary_key: None,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
v1::update::UpdateType::DocumentsPartial { .. } => {
|
||||||
|
v2::updates::UpdateMeta::DocumentsAddition {
|
||||||
|
method: v2::updates::IndexDocumentsMethod::UpdateDocuments,
|
||||||
|
format: v2::updates::UpdateFormat::Json,
|
||||||
|
primary_key: None,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
v1::update::UpdateType::DocumentsDeletion { .. } => {
|
||||||
|
v2::updates::UpdateMeta::DeleteDocuments { ids: vec![] }
|
||||||
|
}
|
||||||
|
v1::update::UpdateType::Settings { settings } => {
|
||||||
|
v2::updates::UpdateMeta::Settings((*settings).into())
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<v1::settings::SettingsUpdate> for v2::Settings<v2::Unchecked> {
|
||||||
|
fn from(source: v1::settings::SettingsUpdate) -> Self {
|
||||||
|
let displayed_attributes: Option<Option<BTreeSet<String>>> =
|
||||||
|
source.displayed_attributes.into();
|
||||||
|
|
||||||
|
let attributes_for_faceting: Option<Option<Vec<String>>> =
|
||||||
|
source.attributes_for_faceting.into();
|
||||||
|
|
||||||
|
let ranking_rules: Option<Option<Vec<v1::settings::RankingRule>>> =
|
||||||
|
source.ranking_rules.into();
|
||||||
|
|
||||||
|
// go from the concrete types of v1 (RankingRule) to the concrete type of v2 (Criterion),
|
||||||
|
// and then back to string as this is what the settings manipulate
|
||||||
|
let ranking_rules = ranking_rules.map(|opt| {
|
||||||
|
opt.map(|ranking_rules| {
|
||||||
|
ranking_rules
|
||||||
|
.into_iter()
|
||||||
|
// filter out the WordsPosition ranking rule that exists in v1 but not v2
|
||||||
|
.filter_map(|ranking_rule| {
|
||||||
|
Option::<v2::settings::Criterion>::from(ranking_rule)
|
||||||
|
})
|
||||||
|
.map(|criterion| criterion.to_string())
|
||||||
|
.collect()
|
||||||
|
})
|
||||||
|
});
|
||||||
|
|
||||||
|
Self {
|
||||||
|
displayed_attributes: displayed_attributes.map(|opt| {
|
||||||
|
opt.map(|displayed_attributes| displayed_attributes.into_iter().collect())
|
||||||
|
}),
|
||||||
|
searchable_attributes: source.searchable_attributes.into(),
|
||||||
|
filterable_attributes: attributes_for_faceting.map(|opt| {
|
||||||
|
opt.map(|attributes_for_faceting| attributes_for_faceting.into_iter().collect())
|
||||||
|
}),
|
||||||
|
ranking_rules,
|
||||||
|
stop_words: source.stop_words.into(),
|
||||||
|
synonyms: source.synonyms.into(),
|
||||||
|
distinct_attribute: source.distinct_attribute.into(),
|
||||||
|
_kind: std::marker::PhantomData,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<v1::settings::RankingRule> for Option<v2::settings::Criterion> {
|
||||||
|
fn from(source: v1::settings::RankingRule) -> Self {
|
||||||
|
match source {
|
||||||
|
v1::settings::RankingRule::Typo => Some(v2::settings::Criterion::Typo),
|
||||||
|
v1::settings::RankingRule::Words => Some(v2::settings::Criterion::Words),
|
||||||
|
v1::settings::RankingRule::Proximity => Some(v2::settings::Criterion::Proximity),
|
||||||
|
v1::settings::RankingRule::Attribute => Some(v2::settings::Criterion::Attribute),
|
||||||
|
v1::settings::RankingRule::WordsPosition => {
|
||||||
|
log::warn!("Removing the 'WordsPosition' ranking rule that is no longer supported, please check the resulting ranking rules of your indexes");
|
||||||
|
None
|
||||||
|
}
|
||||||
|
v1::settings::RankingRule::Exactness => Some(v2::settings::Criterion::Exactness),
|
||||||
|
v1::settings::RankingRule::Asc(field_name) => {
|
||||||
|
Some(v2::settings::Criterion::Asc(field_name))
|
||||||
|
}
|
||||||
|
v1::settings::RankingRule::Desc(field_name) => {
|
||||||
|
Some(v2::settings::Criterion::Desc(field_name))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T> From<v1::settings::UpdateState<T>> for Option<Option<T>> {
|
||||||
|
fn from(source: v1::settings::UpdateState<T>) -> Self {
|
||||||
|
match source {
|
||||||
|
v1::settings::UpdateState::Update(new_value) => Some(Some(new_value)),
|
||||||
|
v1::settings::UpdateState::Clear => Some(None),
|
||||||
|
v1::settings::UpdateState::Nothing => None,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
pub(crate) mod test {
|
||||||
|
use std::fs::File;
|
||||||
|
use std::io::BufReader;
|
||||||
|
|
||||||
|
use flate2::bufread::GzDecoder;
|
||||||
|
use meili_snap::insta;
|
||||||
|
use tempfile::TempDir;
|
||||||
|
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn compat_v1_v2() {
|
||||||
|
let dump = File::open("tests/assets/v1.dump").unwrap();
|
||||||
|
let dir = TempDir::new().unwrap();
|
||||||
|
let mut dump = BufReader::new(dump);
|
||||||
|
let gz = GzDecoder::new(&mut dump);
|
||||||
|
let mut archive = tar::Archive::new(gz);
|
||||||
|
archive.unpack(dir.path()).unwrap();
|
||||||
|
|
||||||
|
let mut dump = v1::V1Reader::open(dir).unwrap().to_v2();
|
||||||
|
|
||||||
|
// top level infos
|
||||||
|
assert_eq!(dump.date(), None);
|
||||||
|
|
||||||
|
// tasks
|
||||||
|
let tasks = dump.tasks().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
let (tasks, update_files): (Vec<_>, Vec<_>) = tasks.into_iter().unzip();
|
||||||
|
meili_snap::snapshot_hash!(meili_snap::json_string!(tasks), @"ad6245d98d1a8e30535f3339a9a8d223");
|
||||||
|
assert_eq!(update_files.len(), 9);
|
||||||
|
assert!(update_files[..].iter().all(|u| u.is_none())); // no update file in dumps v1
|
||||||
|
|
||||||
|
// indexes
|
||||||
|
let mut indexes = dump.indexes().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
// the index are not ordered in any way by default
|
||||||
|
indexes.sort_by_key(|index| index.metadata().uid.to_string());
|
||||||
|
|
||||||
|
let mut products = indexes.pop().unwrap();
|
||||||
|
let mut movies = indexes.pop().unwrap();
|
||||||
|
let mut spells = indexes.pop().unwrap();
|
||||||
|
assert!(indexes.is_empty());
|
||||||
|
|
||||||
|
// products
|
||||||
|
insta::assert_json_snapshot!(products.metadata(), { ".createdAt" => "[now]", ".updatedAt" => "[now]" }, @r###"
|
||||||
|
{
|
||||||
|
"uid": "products",
|
||||||
|
"primaryKey": "sku",
|
||||||
|
"createdAt": "[now]",
|
||||||
|
"updatedAt": "[now]"
|
||||||
|
}
|
||||||
|
"###);
|
||||||
|
|
||||||
|
insta::assert_json_snapshot!(products.settings().unwrap());
|
||||||
|
let documents = products.documents().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
assert_eq!(documents.len(), 10);
|
||||||
|
meili_snap::snapshot_hash!(format!("{:#?}", documents), @"b01c8371aea4c7171af0d4d846a2bdca");
|
||||||
|
|
||||||
|
// movies
|
||||||
|
insta::assert_json_snapshot!(movies.metadata(), { ".createdAt" => "[now]", ".updatedAt" => "[now]" }, @r###"
|
||||||
|
{
|
||||||
|
"uid": "movies",
|
||||||
|
"primaryKey": "id",
|
||||||
|
"createdAt": "[now]",
|
||||||
|
"updatedAt": "[now]"
|
||||||
|
}
|
||||||
|
"###);
|
||||||
|
|
||||||
|
insta::assert_json_snapshot!(movies.settings().unwrap());
|
||||||
|
let documents = movies.documents().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
assert_eq!(documents.len(), 10);
|
||||||
|
meili_snap::snapshot_hash!(format!("{:#?}", documents), @"b63dbed5bbc059f3e32bc471ae699bf5");
|
||||||
|
|
||||||
|
// spells
|
||||||
|
insta::assert_json_snapshot!(spells.metadata(), { ".createdAt" => "[now]", ".updatedAt" => "[now]" }, @r###"
|
||||||
|
{
|
||||||
|
"uid": "dnd_spells",
|
||||||
|
"primaryKey": "index",
|
||||||
|
"createdAt": "[now]",
|
||||||
|
"updatedAt": "[now]"
|
||||||
|
}
|
||||||
|
"###);
|
||||||
|
|
||||||
|
insta::assert_json_snapshot!(spells.settings().unwrap());
|
||||||
|
let documents = spells.documents().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
assert_eq!(documents.len(), 10);
|
||||||
|
meili_snap::snapshot_hash!(format!("{:#?}", documents), @"aa24c0cfc733d66c396237ad44263bed");
|
||||||
|
}
|
||||||
|
}
|
@ -4,22 +4,28 @@ use std::str::FromStr;
|
|||||||
use time::OffsetDateTime;
|
use time::OffsetDateTime;
|
||||||
use uuid::Uuid;
|
use uuid::Uuid;
|
||||||
|
|
||||||
|
use super::v1_to_v2::{CompatIndexV1ToV2, CompatV1ToV2};
|
||||||
use super::v3_to_v4::CompatV3ToV4;
|
use super::v3_to_v4::CompatV3ToV4;
|
||||||
use crate::reader::{v2, v3, Document};
|
use crate::reader::{v2, v3, Document};
|
||||||
use crate::Result;
|
use crate::Result;
|
||||||
|
|
||||||
pub struct CompatV2ToV3 {
|
pub enum CompatV2ToV3 {
|
||||||
pub from: v2::V2Reader,
|
V2(v2::V2Reader),
|
||||||
|
Compat(CompatV1ToV2),
|
||||||
}
|
}
|
||||||
|
|
||||||
impl CompatV2ToV3 {
|
impl CompatV2ToV3 {
|
||||||
pub fn new(v2: v2::V2Reader) -> CompatV2ToV3 {
|
pub fn new(v2: v2::V2Reader) -> CompatV2ToV3 {
|
||||||
CompatV2ToV3 { from: v2 }
|
CompatV2ToV3::V2(v2)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn index_uuid(&self) -> Vec<v3::meta::IndexUuid> {
|
pub fn index_uuid(&self) -> Vec<v3::meta::IndexUuid> {
|
||||||
self.from
|
let v2_uuids = match self {
|
||||||
.index_uuid()
|
CompatV2ToV3::V2(from) => from.index_uuid(),
|
||||||
|
CompatV2ToV3::Compat(compat) => compat.index_uuid(),
|
||||||
|
};
|
||||||
|
v2_uuids
|
||||||
|
.into_iter()
|
||||||
.into_iter()
|
.into_iter()
|
||||||
.map(|index| v3::meta::IndexUuid { uid: index.uid, uuid: index.uuid })
|
.map(|index| v3::meta::IndexUuid { uid: index.uid, uuid: index.uuid })
|
||||||
.collect()
|
.collect()
|
||||||
@ -30,11 +36,17 @@ impl CompatV2ToV3 {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn version(&self) -> crate::Version {
|
pub fn version(&self) -> crate::Version {
|
||||||
self.from.version()
|
match self {
|
||||||
|
CompatV2ToV3::V2(from) => from.version(),
|
||||||
|
CompatV2ToV3::Compat(compat) => compat.version(),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn date(&self) -> Option<time::OffsetDateTime> {
|
pub fn date(&self) -> Option<time::OffsetDateTime> {
|
||||||
self.from.date()
|
match self {
|
||||||
|
CompatV2ToV3::V2(from) => from.date(),
|
||||||
|
CompatV2ToV3::Compat(compat) => compat.date(),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn instance_uid(&self) -> Result<Option<uuid::Uuid>> {
|
pub fn instance_uid(&self) -> Result<Option<uuid::Uuid>> {
|
||||||
@ -42,10 +54,18 @@ impl CompatV2ToV3 {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn indexes(&self) -> Result<impl Iterator<Item = Result<CompatIndexV2ToV3>> + '_> {
|
pub fn indexes(&self) -> Result<impl Iterator<Item = Result<CompatIndexV2ToV3>> + '_> {
|
||||||
Ok(self.from.indexes()?.map(|index_reader| -> Result<_> {
|
Ok(match self {
|
||||||
let compat = CompatIndexV2ToV3::new(index_reader?);
|
CompatV2ToV3::V2(from) => Box::new(from.indexes()?.map(|index_reader| -> Result<_> {
|
||||||
Ok(compat)
|
let compat = CompatIndexV2ToV3::new(index_reader?);
|
||||||
}))
|
Ok(compat)
|
||||||
|
}))
|
||||||
|
as Box<dyn Iterator<Item = Result<CompatIndexV2ToV3>> + '_>,
|
||||||
|
CompatV2ToV3::Compat(compat) => Box::new(compat.indexes()?.map(|index_reader| {
|
||||||
|
let compat = CompatIndexV2ToV3::Compat(Box::new(index_reader?));
|
||||||
|
Ok(compat)
|
||||||
|
}))
|
||||||
|
as Box<dyn Iterator<Item = Result<CompatIndexV2ToV3>> + '_>,
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn tasks(
|
pub fn tasks(
|
||||||
@ -54,11 +74,13 @@ impl CompatV2ToV3 {
|
|||||||
dyn Iterator<Item = Result<(v3::Task, Option<Box<dyn Iterator<Item = Result<Document>>>>)>>
|
dyn Iterator<Item = Result<(v3::Task, Option<Box<dyn Iterator<Item = Result<Document>>>>)>>
|
||||||
+ '_,
|
+ '_,
|
||||||
> {
|
> {
|
||||||
let _indexes = self.from.index_uuid.clone();
|
let tasks = match self {
|
||||||
|
CompatV2ToV3::V2(from) => from.tasks(),
|
||||||
|
CompatV2ToV3::Compat(compat) => compat.tasks(),
|
||||||
|
};
|
||||||
|
|
||||||
Box::new(
|
Box::new(
|
||||||
self.from
|
tasks
|
||||||
.tasks()
|
|
||||||
.map(move |task| {
|
.map(move |task| {
|
||||||
task.map(|(task, content_file)| {
|
task.map(|(task, content_file)| {
|
||||||
let task = v3::Task { uuid: task.uuid, update: task.update.into() };
|
let task = v3::Task { uuid: task.uuid, update: task.update.into() };
|
||||||
@ -76,27 +98,38 @@ impl CompatV2ToV3 {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub struct CompatIndexV2ToV3 {
|
pub enum CompatIndexV2ToV3 {
|
||||||
from: v2::V2IndexReader,
|
V2(v2::V2IndexReader),
|
||||||
|
Compat(Box<CompatIndexV1ToV2>),
|
||||||
}
|
}
|
||||||
|
|
||||||
impl CompatIndexV2ToV3 {
|
impl CompatIndexV2ToV3 {
|
||||||
pub fn new(v2: v2::V2IndexReader) -> CompatIndexV2ToV3 {
|
pub fn new(v2: v2::V2IndexReader) -> CompatIndexV2ToV3 {
|
||||||
CompatIndexV2ToV3 { from: v2 }
|
CompatIndexV2ToV3::V2(v2)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn metadata(&self) -> &crate::IndexMetadata {
|
pub fn metadata(&self) -> &crate::IndexMetadata {
|
||||||
self.from.metadata()
|
match self {
|
||||||
|
CompatIndexV2ToV3::V2(from) => from.metadata(),
|
||||||
|
CompatIndexV2ToV3::Compat(compat) => compat.metadata(),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn documents(&mut self) -> Result<Box<dyn Iterator<Item = Result<Document>> + '_>> {
|
pub fn documents(&mut self) -> Result<Box<dyn Iterator<Item = Result<Document>> + '_>> {
|
||||||
self.from
|
match self {
|
||||||
.documents()
|
CompatIndexV2ToV3::V2(from) => from
|
||||||
.map(|iter| Box::new(iter) as Box<dyn Iterator<Item = Result<Document>> + '_>)
|
.documents()
|
||||||
|
.map(|iter| Box::new(iter) as Box<dyn Iterator<Item = Result<Document>> + '_>),
|
||||||
|
CompatIndexV2ToV3::Compat(compat) => compat.documents(),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn settings(&mut self) -> Result<v3::Settings<v3::Checked>> {
|
pub fn settings(&mut self) -> Result<v3::Settings<v3::Checked>> {
|
||||||
Ok(v3::Settings::<v3::Unchecked>::from(self.from.settings()?).check())
|
let settings = match self {
|
||||||
|
CompatIndexV2ToV3::V2(from) => from.settings()?,
|
||||||
|
CompatIndexV2ToV3::Compat(compat) => compat.settings()?,
|
||||||
|
};
|
||||||
|
Ok(v3::Settings::<v3::Unchecked>::from(settings).check())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -9,11 +9,11 @@ use self::compat::v4_to_v5::CompatV4ToV5;
|
|||||||
use self::compat::v5_to_v6::{CompatIndexV5ToV6, CompatV5ToV6};
|
use self::compat::v5_to_v6::{CompatIndexV5ToV6, CompatV5ToV6};
|
||||||
use self::v5::V5Reader;
|
use self::v5::V5Reader;
|
||||||
use self::v6::{V6IndexReader, V6Reader};
|
use self::v6::{V6IndexReader, V6Reader};
|
||||||
use crate::{Error, Result, Version};
|
use crate::{Result, Version};
|
||||||
|
|
||||||
mod compat;
|
mod compat;
|
||||||
|
|
||||||
// pub(self) mod v1;
|
pub(self) mod v1;
|
||||||
pub(self) mod v2;
|
pub(self) mod v2;
|
||||||
pub(self) mod v3;
|
pub(self) mod v3;
|
||||||
pub(self) mod v4;
|
pub(self) mod v4;
|
||||||
@ -45,8 +45,9 @@ impl DumpReader {
|
|||||||
let MetadataVersion { dump_version } = serde_json::from_reader(&mut meta_file)?;
|
let MetadataVersion { dump_version } = serde_json::from_reader(&mut meta_file)?;
|
||||||
|
|
||||||
match dump_version {
|
match dump_version {
|
||||||
// Version::V1 => Ok(Box::new(v1::Reader::open(path)?)),
|
Version::V1 => {
|
||||||
Version::V1 => Err(Error::DumpV1Unsupported),
|
Ok(v1::V1Reader::open(path)?.to_v2().to_v3().to_v4().to_v5().to_v6().into())
|
||||||
|
}
|
||||||
Version::V2 => Ok(v2::V2Reader::open(path)?.to_v3().to_v4().to_v5().to_v6().into()),
|
Version::V2 => Ok(v2::V2Reader::open(path)?.to_v3().to_v4().to_v5().to_v6().into()),
|
||||||
Version::V3 => Ok(v3::V3Reader::open(path)?.to_v4().to_v5().to_v6().into()),
|
Version::V3 => Ok(v3::V3Reader::open(path)?.to_v4().to_v5().to_v6().into()),
|
||||||
Version::V4 => Ok(v4::V4Reader::open(path)?.to_v5().to_v6().into()),
|
Version::V4 => Ok(v4::V4Reader::open(path)?.to_v5().to_v6().into()),
|
||||||
@ -528,4 +529,81 @@ pub(crate) mod test {
|
|||||||
assert_eq!(documents.len(), 10);
|
assert_eq!(documents.len(), 10);
|
||||||
meili_snap::snapshot_hash!(format!("{:#?}", documents), @"235016433dd04262c7f2da01d1e808ce");
|
meili_snap::snapshot_hash!(format!("{:#?}", documents), @"235016433dd04262c7f2da01d1e808ce");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn import_dump_v1() {
|
||||||
|
let dump = File::open("tests/assets/v1.dump").unwrap();
|
||||||
|
let mut dump = DumpReader::open(dump).unwrap();
|
||||||
|
|
||||||
|
// top level infos
|
||||||
|
assert_eq!(dump.date(), None);
|
||||||
|
assert_eq!(dump.instance_uid().unwrap(), None);
|
||||||
|
|
||||||
|
// tasks
|
||||||
|
let tasks = dump.tasks().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
let (tasks, update_files): (Vec<_>, Vec<_>) = tasks.into_iter().unzip();
|
||||||
|
meili_snap::snapshot_hash!(meili_snap::json_string!(tasks), @"b3e3652bfc10a76670be157d2507d761");
|
||||||
|
assert_eq!(update_files.len(), 9);
|
||||||
|
assert!(update_files[..].iter().all(|u| u.is_none())); // no update file in dump v1
|
||||||
|
|
||||||
|
// keys
|
||||||
|
let keys = dump.keys().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
meili_snap::snapshot!(meili_snap::json_string!(keys), @"[]");
|
||||||
|
meili_snap::snapshot_hash!(meili_snap::json_string!(keys), @"d751713988987e9331980363e24189ce");
|
||||||
|
|
||||||
|
// indexes
|
||||||
|
let mut indexes = dump.indexes().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
// the index are not ordered in any way by default
|
||||||
|
indexes.sort_by_key(|index| index.metadata().uid.to_string());
|
||||||
|
|
||||||
|
let mut products = indexes.pop().unwrap();
|
||||||
|
let mut movies = indexes.pop().unwrap();
|
||||||
|
let mut spells = indexes.pop().unwrap();
|
||||||
|
assert!(indexes.is_empty());
|
||||||
|
|
||||||
|
// products
|
||||||
|
insta::assert_json_snapshot!(products.metadata(), { ".createdAt" => "[now]", ".updatedAt" => "[now]" }, @r###"
|
||||||
|
{
|
||||||
|
"uid": "products",
|
||||||
|
"primaryKey": "sku",
|
||||||
|
"createdAt": "[now]",
|
||||||
|
"updatedAt": "[now]"
|
||||||
|
}
|
||||||
|
"###);
|
||||||
|
|
||||||
|
insta::assert_json_snapshot!(products.settings().unwrap());
|
||||||
|
let documents = products.documents().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
assert_eq!(documents.len(), 10);
|
||||||
|
meili_snap::snapshot_hash!(format!("{:#?}", documents), @"b01c8371aea4c7171af0d4d846a2bdca");
|
||||||
|
|
||||||
|
// movies
|
||||||
|
insta::assert_json_snapshot!(movies.metadata(), { ".createdAt" => "[now]", ".updatedAt" => "[now]" }, @r###"
|
||||||
|
{
|
||||||
|
"uid": "movies",
|
||||||
|
"primaryKey": "id",
|
||||||
|
"createdAt": "[now]",
|
||||||
|
"updatedAt": "[now]"
|
||||||
|
}
|
||||||
|
"###);
|
||||||
|
|
||||||
|
insta::assert_json_snapshot!(movies.settings().unwrap());
|
||||||
|
let documents = movies.documents().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
assert_eq!(documents.len(), 10);
|
||||||
|
meili_snap::snapshot_hash!(format!("{:#?}", documents), @"b63dbed5bbc059f3e32bc471ae699bf5");
|
||||||
|
|
||||||
|
// spells
|
||||||
|
insta::assert_json_snapshot!(spells.metadata(), { ".createdAt" => "[now]", ".updatedAt" => "[now]" }, @r###"
|
||||||
|
{
|
||||||
|
"uid": "dnd_spells",
|
||||||
|
"primaryKey": "index",
|
||||||
|
"createdAt": "[now]",
|
||||||
|
"updatedAt": "[now]"
|
||||||
|
}
|
||||||
|
"###);
|
||||||
|
|
||||||
|
insta::assert_json_snapshot!(spells.settings().unwrap());
|
||||||
|
let documents = spells.documents().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
assert_eq!(documents.len(), 10);
|
||||||
|
meili_snap::snapshot_hash!(format!("{:#?}", documents), @"aa24c0cfc733d66c396237ad44263bed");
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@ -0,0 +1,27 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/mod.rs
|
||||||
|
expression: movies.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"filterableAttributes": [
|
||||||
|
"genres",
|
||||||
|
"id"
|
||||||
|
],
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"exactness",
|
||||||
|
"release_date:asc"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"distinctAttribute": null
|
||||||
|
}
|
@ -0,0 +1,23 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/mod.rs
|
||||||
|
expression: spells.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"filterableAttributes": [],
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"exactness"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"distinctAttribute": null
|
||||||
|
}
|
@ -0,0 +1,23 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/mod.rs
|
||||||
|
expression: spells.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"filterableAttributes": [],
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"exactness"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"distinctAttribute": null
|
||||||
|
}
|
@ -0,0 +1,37 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/mod.rs
|
||||||
|
expression: products.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"filterableAttributes": [],
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"exactness"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {
|
||||||
|
"android": [
|
||||||
|
"phone",
|
||||||
|
"smartphone"
|
||||||
|
],
|
||||||
|
"iphone": [
|
||||||
|
"phone",
|
||||||
|
"smartphone"
|
||||||
|
],
|
||||||
|
"phone": [
|
||||||
|
"android",
|
||||||
|
"iphone",
|
||||||
|
"smartphone"
|
||||||
|
]
|
||||||
|
},
|
||||||
|
"distinctAttribute": null
|
||||||
|
}
|
@ -0,0 +1,37 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/mod.rs
|
||||||
|
expression: products.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"filterableAttributes": [],
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"exactness"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {
|
||||||
|
"android": [
|
||||||
|
"phone",
|
||||||
|
"smartphone"
|
||||||
|
],
|
||||||
|
"iphone": [
|
||||||
|
"phone",
|
||||||
|
"smartphone"
|
||||||
|
],
|
||||||
|
"phone": [
|
||||||
|
"android",
|
||||||
|
"iphone",
|
||||||
|
"smartphone"
|
||||||
|
]
|
||||||
|
},
|
||||||
|
"distinctAttribute": null
|
||||||
|
}
|
@ -0,0 +1,27 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/mod.rs
|
||||||
|
expression: movies.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"filterableAttributes": [
|
||||||
|
"genres",
|
||||||
|
"id"
|
||||||
|
],
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"exactness",
|
||||||
|
"release_date:asc"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"distinctAttribute": null
|
||||||
|
}
|
@ -1,173 +1,263 @@
|
|||||||
use std::{
|
use std::{
|
||||||
convert::Infallible,
|
|
||||||
fs::{self, File},
|
fs::{self, File},
|
||||||
io::{BufRead, BufReader},
|
io::{BufRead, BufReader},
|
||||||
path::Path,
|
path::{Path, PathBuf},
|
||||||
};
|
};
|
||||||
|
|
||||||
use tempfile::TempDir;
|
use tempfile::TempDir;
|
||||||
use time::OffsetDateTime;
|
use time::OffsetDateTime;
|
||||||
|
|
||||||
use self::update::UpdateStatus;
|
use super::{compat::v1_to_v2::CompatV1ToV2, Document};
|
||||||
|
use crate::{IndexMetadata, Result, Version};
|
||||||
use super::{DumpReader, IndexReader};
|
use serde::Deserialize;
|
||||||
use crate::{Error, Result, Version};
|
|
||||||
|
|
||||||
pub mod settings;
|
pub mod settings;
|
||||||
pub mod update;
|
pub mod update;
|
||||||
pub mod v1;
|
|
||||||
|
|
||||||
pub struct V1Reader {
|
pub struct V1Reader {
|
||||||
dump: TempDir,
|
pub dump: TempDir,
|
||||||
metadata: v1::Metadata,
|
pub db_version: String,
|
||||||
indexes: Vec<V1IndexReader>,
|
pub dump_version: crate::Version,
|
||||||
|
indexes: Vec<V1Index>,
|
||||||
}
|
}
|
||||||
|
|
||||||
struct V1IndexReader {
|
pub struct IndexUuid {
|
||||||
name: String,
|
pub name: String,
|
||||||
|
pub uid: String,
|
||||||
|
}
|
||||||
|
pub type Task = self::update::UpdateStatus;
|
||||||
|
|
||||||
|
struct V1Index {
|
||||||
|
metadata: IndexMetadataV1,
|
||||||
|
path: PathBuf,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl V1Index {
|
||||||
|
pub fn new(path: PathBuf, metadata: Index) -> Self {
|
||||||
|
Self { metadata: metadata.into(), path }
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn open(&self) -> Result<V1IndexReader> {
|
||||||
|
V1IndexReader::new(&self.path, self.metadata.clone())
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn metadata(&self) -> &IndexMetadata {
|
||||||
|
&self.metadata.metadata
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct V1IndexReader {
|
||||||
|
metadata: IndexMetadataV1,
|
||||||
documents: BufReader<File>,
|
documents: BufReader<File>,
|
||||||
settings: BufReader<File>,
|
settings: BufReader<File>,
|
||||||
updates: BufReader<File>,
|
updates: BufReader<File>,
|
||||||
|
|
||||||
current_update: Option<UpdateStatus>,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl V1IndexReader {
|
impl V1IndexReader {
|
||||||
pub fn new(name: String, path: &Path) -> Result<Self> {
|
pub fn new(path: &Path, metadata: IndexMetadataV1) -> Result<Self> {
|
||||||
let mut ret = V1IndexReader {
|
Ok(V1IndexReader {
|
||||||
name,
|
metadata,
|
||||||
documents: BufReader::new(File::open(path.join("documents.jsonl"))?),
|
documents: BufReader::new(File::open(path.join("documents.jsonl"))?),
|
||||||
settings: BufReader::new(File::open(path.join("settings.json"))?),
|
settings: BufReader::new(File::open(path.join("settings.json"))?),
|
||||||
updates: BufReader::new(File::open(path.join("updates.jsonl"))?),
|
updates: BufReader::new(File::open(path.join("updates.jsonl"))?),
|
||||||
current_update: None,
|
})
|
||||||
};
|
|
||||||
ret.next_update();
|
|
||||||
|
|
||||||
Ok(ret)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn next_update(&mut self) -> Result<Option<UpdateStatus>> {
|
pub fn metadata(&self) -> &IndexMetadata {
|
||||||
let current_update = if let Some(line) = self.updates.lines().next() {
|
&self.metadata.metadata
|
||||||
Some(serde_json::from_str(&line?)?)
|
}
|
||||||
} else {
|
|
||||||
None
|
|
||||||
};
|
|
||||||
|
|
||||||
Ok(std::mem::replace(&mut self.current_update, current_update))
|
pub fn documents(&mut self) -> Result<impl Iterator<Item = Result<Document>> + '_> {
|
||||||
|
Ok((&mut self.documents)
|
||||||
|
.lines()
|
||||||
|
.map(|line| -> Result<_> { Ok(serde_json::from_str(&line?)?) }))
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn settings(&mut self) -> Result<self::settings::Settings> {
|
||||||
|
Ok(serde_json::from_reader(&mut self.settings)?)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn tasks(self) -> impl Iterator<Item = Result<Task>> {
|
||||||
|
self.updates.lines().map(|line| -> Result<_> { Ok(serde_json::from_str(&line?)?) })
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl V1Reader {
|
impl V1Reader {
|
||||||
pub fn open(dump: TempDir) -> Result<Self> {
|
pub fn open(dump: TempDir) -> Result<Self> {
|
||||||
let mut meta_file = fs::read(dump.path().join("metadata.json"))?;
|
let meta_file = fs::read(dump.path().join("metadata.json"))?;
|
||||||
let metadata = serde_json::from_reader(&*meta_file)?;
|
let metadata: Metadata = serde_json::from_reader(&*meta_file)?;
|
||||||
|
|
||||||
let mut indexes = Vec::new();
|
let mut indexes = Vec::new();
|
||||||
|
|
||||||
let entries = fs::read_dir(dump.path())?;
|
for index in metadata.indexes.into_iter() {
|
||||||
for entry in entries {
|
let index_path = dump.path().join(&index.uid);
|
||||||
let entry = entry?;
|
indexes.push(V1Index::new(index_path, index));
|
||||||
if entry.file_type()?.is_dir() {
|
|
||||||
indexes.push(V1IndexReader::new(
|
|
||||||
entry
|
|
||||||
.file_name()
|
|
||||||
.to_str()
|
|
||||||
.ok_or(Error::BadIndexName)?
|
|
||||||
.to_string(),
|
|
||||||
&entry.path(),
|
|
||||||
)?);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(V1Reader {
|
Ok(V1Reader {
|
||||||
dump,
|
dump,
|
||||||
metadata,
|
|
||||||
indexes,
|
indexes,
|
||||||
|
db_version: metadata.db_version,
|
||||||
|
dump_version: metadata.dump_version,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
fn next_update(&mut self) -> Result<Option<UpdateStatus>> {
|
pub fn to_v2(self) -> CompatV1ToV2 {
|
||||||
if let Some((idx, _)) = self
|
CompatV1ToV2 { from: self }
|
||||||
.indexes
|
}
|
||||||
|
|
||||||
|
pub fn index_uuid(&self) -> Vec<IndexUuid> {
|
||||||
|
self.indexes
|
||||||
.iter()
|
.iter()
|
||||||
.map(|index| index.current_update)
|
.map(|index| IndexUuid {
|
||||||
.enumerate()
|
name: index.metadata.name.to_owned(),
|
||||||
.filter_map(|(idx, update)| update.map(|u| (idx, u)))
|
uid: index.metadata().uid.to_owned(),
|
||||||
.min_by_key(|(_, update)| update.enqueued_at())
|
})
|
||||||
{
|
.collect()
|
||||||
self.indexes[idx].next_update()
|
}
|
||||||
} else {
|
|
||||||
Ok(None)
|
pub fn version(&self) -> Version {
|
||||||
|
Version::V1
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn date(&self) -> Option<OffsetDateTime> {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn indexes(&self) -> Result<impl Iterator<Item = Result<V1IndexReader>> + '_> {
|
||||||
|
Ok(self.indexes.iter().map(|index| index.open()))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Deserialize)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct Index {
|
||||||
|
pub name: String,
|
||||||
|
pub uid: String,
|
||||||
|
#[serde(with = "time::serde::rfc3339")]
|
||||||
|
created_at: OffsetDateTime,
|
||||||
|
#[serde(with = "time::serde::rfc3339")]
|
||||||
|
updated_at: OffsetDateTime,
|
||||||
|
pub primary_key: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Clone)]
|
||||||
|
pub struct IndexMetadataV1 {
|
||||||
|
pub name: String,
|
||||||
|
pub metadata: crate::IndexMetadata,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<Index> for IndexMetadataV1 {
|
||||||
|
fn from(index: Index) -> Self {
|
||||||
|
IndexMetadataV1 {
|
||||||
|
name: index.name,
|
||||||
|
metadata: crate::IndexMetadata {
|
||||||
|
uid: index.uid,
|
||||||
|
primary_key: index.primary_key,
|
||||||
|
created_at: index.created_at,
|
||||||
|
updated_at: index.updated_at,
|
||||||
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl IndexReader for &V1IndexReader {
|
#[derive(Debug, Deserialize)]
|
||||||
type Document = serde_json::Map<String, serde_json::Value>;
|
#[serde(rename_all = "camelCase")]
|
||||||
type Settings = settings::Settings;
|
pub struct Metadata {
|
||||||
|
pub indexes: Vec<Index>,
|
||||||
fn name(&self) -> &str {
|
pub db_version: String,
|
||||||
todo!()
|
pub dump_version: crate::Version,
|
||||||
}
|
|
||||||
|
|
||||||
fn documents(&self) -> Result<Box<dyn Iterator<Item = Result<Self::Document>>>> {
|
|
||||||
todo!()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn settings(&self) -> Result<Self::Settings> {
|
|
||||||
todo!()
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl DumpReader for V1Reader {
|
#[cfg(test)]
|
||||||
type Document = serde_json::Map<String, serde_json::Value>;
|
pub(crate) mod test {
|
||||||
type Settings = settings::Settings;
|
use std::fs::File;
|
||||||
|
use std::io::BufReader;
|
||||||
|
|
||||||
type Task = update::UpdateStatus;
|
use flate2::bufread::GzDecoder;
|
||||||
type UpdateFile = Infallible;
|
use meili_snap::insta;
|
||||||
|
use tempfile::TempDir;
|
||||||
|
|
||||||
type Key = Infallible;
|
use super::*;
|
||||||
|
|
||||||
fn date(&self) -> Option<OffsetDateTime> {
|
#[test]
|
||||||
None
|
fn read_dump_v1() {
|
||||||
}
|
let dump = File::open("tests/assets/v1.dump").unwrap();
|
||||||
|
let dir = TempDir::new().unwrap();
|
||||||
|
let mut dump = BufReader::new(dump);
|
||||||
|
let gz = GzDecoder::new(&mut dump);
|
||||||
|
let mut archive = tar::Archive::new(gz);
|
||||||
|
archive.unpack(dir.path()).unwrap();
|
||||||
|
|
||||||
fn version(&self) -> Version {
|
let dump = V1Reader::open(dir).unwrap();
|
||||||
Version::V1
|
|
||||||
}
|
|
||||||
|
|
||||||
fn indexes(
|
// top level infos
|
||||||
&self,
|
assert_eq!(dump.date(), None);
|
||||||
) -> Result<
|
|
||||||
Box<
|
|
||||||
dyn Iterator<
|
|
||||||
Item = Result<
|
|
||||||
Box<
|
|
||||||
dyn super::IndexReader<
|
|
||||||
Document = Self::Document,
|
|
||||||
Settings = Self::Settings,
|
|
||||||
>,
|
|
||||||
>,
|
|
||||||
>,
|
|
||||||
>,
|
|
||||||
>,
|
|
||||||
> {
|
|
||||||
Ok(Box::new(self.indexes.iter().map(|index| {
|
|
||||||
let index = Box::new(index)
|
|
||||||
as Box<dyn IndexReader<Document = Self::Document, Settings = Self::Settings>>;
|
|
||||||
Ok(index)
|
|
||||||
})))
|
|
||||||
}
|
|
||||||
|
|
||||||
fn tasks(&self) -> Box<dyn Iterator<Item = Result<(Self::Task, Option<Self::UpdateFile>)>>> {
|
// indexes
|
||||||
Box::new(std::iter::from_fn(|| {
|
let mut indexes = dump.indexes().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
self.next_update()
|
|
||||||
.transpose()
|
|
||||||
.map(|result| result.map(|task| (task, None)))
|
|
||||||
}))
|
|
||||||
}
|
|
||||||
|
|
||||||
fn keys(&self) -> Box<dyn Iterator<Item = Result<Self::Key>>> {
|
let mut products = indexes.pop().unwrap();
|
||||||
Box::new(std::iter::empty())
|
let mut movies = indexes.pop().unwrap();
|
||||||
|
let mut dnd_spells = indexes.pop().unwrap();
|
||||||
|
|
||||||
|
assert!(indexes.is_empty());
|
||||||
|
|
||||||
|
// products
|
||||||
|
insta::assert_json_snapshot!(products.metadata(), { ".createdAt" => "[now]", ".updatedAt" => "[now]" }, @r###"
|
||||||
|
{
|
||||||
|
"uid": "products",
|
||||||
|
"primaryKey": "sku",
|
||||||
|
"createdAt": "[now]",
|
||||||
|
"updatedAt": "[now]"
|
||||||
|
}
|
||||||
|
"###);
|
||||||
|
|
||||||
|
insta::assert_json_snapshot!(products.settings().unwrap());
|
||||||
|
let documents = products.documents().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
assert_eq!(documents.len(), 10);
|
||||||
|
meili_snap::snapshot_hash!(format!("{:#?}", documents), @"b01c8371aea4c7171af0d4d846a2bdca");
|
||||||
|
|
||||||
|
// products tasks
|
||||||
|
let tasks = products.tasks().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
meili_snap::snapshot_hash!(meili_snap::json_string!(tasks), @"91de507f206ad21964584021932ba7a7");
|
||||||
|
|
||||||
|
// movies
|
||||||
|
insta::assert_json_snapshot!(movies.metadata(), { ".createdAt" => "[now]", ".updatedAt" => "[now]" }, @r###"
|
||||||
|
{
|
||||||
|
"uid": "movies",
|
||||||
|
"primaryKey": "id",
|
||||||
|
"createdAt": "[now]",
|
||||||
|
"updatedAt": "[now]"
|
||||||
|
}
|
||||||
|
"###);
|
||||||
|
|
||||||
|
insta::assert_json_snapshot!(movies.settings().unwrap());
|
||||||
|
let documents = movies.documents().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
assert_eq!(documents.len(), 10);
|
||||||
|
meili_snap::snapshot_hash!(format!("{:#?}", documents), @"b63dbed5bbc059f3e32bc471ae699bf5");
|
||||||
|
|
||||||
|
// movies tasks
|
||||||
|
let tasks = movies.tasks().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
meili_snap::snapshot_hash!(meili_snap::json_string!(tasks), @"55eef4de2bef7e84c5ce0bee47488f56");
|
||||||
|
|
||||||
|
// spells
|
||||||
|
insta::assert_json_snapshot!(dnd_spells.metadata(), { ".createdAt" => "[now]", ".updatedAt" => "[now]" }, @r###"
|
||||||
|
{
|
||||||
|
"uid": "dnd_spells",
|
||||||
|
"primaryKey": "index",
|
||||||
|
"createdAt": "[now]",
|
||||||
|
"updatedAt": "[now]"
|
||||||
|
}
|
||||||
|
"###);
|
||||||
|
|
||||||
|
insta::assert_json_snapshot!(dnd_spells.settings().unwrap());
|
||||||
|
let documents = dnd_spells.documents().unwrap().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
assert_eq!(documents.len(), 10);
|
||||||
|
meili_snap::snapshot_hash!(format!("{:#?}", documents), @"aa24c0cfc733d66c396237ad44263bed");
|
||||||
|
|
||||||
|
// spells tasks
|
||||||
|
let tasks = dnd_spells.tasks().collect::<Result<Vec<_>>>().unwrap();
|
||||||
|
meili_snap::snapshot_hash!(meili_snap::json_string!(tasks), @"836dd7d64d5ad20ad901c44b1b161a4c");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1,6 +1,9 @@
|
|||||||
use std::collections::{BTreeMap, BTreeSet};
|
use std::collections::{BTreeMap, BTreeSet};
|
||||||
use std::result::Result as StdResult;
|
use std::result::Result as StdResult;
|
||||||
|
use std::str::FromStr;
|
||||||
|
|
||||||
|
use once_cell::sync::Lazy;
|
||||||
|
use regex::Regex;
|
||||||
use serde::{Deserialize, Deserializer, Serialize};
|
use serde::{Deserialize, Deserializer, Serialize};
|
||||||
|
|
||||||
#[derive(Default, Clone, Serialize, Deserialize, Debug)]
|
#[derive(Default, Clone, Serialize, Deserialize, Debug)]
|
||||||
@ -53,6 +56,34 @@ pub enum RankingRule {
|
|||||||
Desc(String),
|
Desc(String),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static ASC_DESC_REGEX: Lazy<Regex> =
|
||||||
|
Lazy::new(|| Regex::new(r#"(asc|desc)\(([\w_-]+)\)"#).unwrap());
|
||||||
|
|
||||||
|
impl FromStr for RankingRule {
|
||||||
|
type Err = ();
|
||||||
|
|
||||||
|
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
||||||
|
Ok(match s {
|
||||||
|
"typo" => Self::Typo,
|
||||||
|
"words" => Self::Words,
|
||||||
|
"proximity" => Self::Proximity,
|
||||||
|
"attribute" => Self::Attribute,
|
||||||
|
"wordsPosition" => Self::WordsPosition,
|
||||||
|
"exactness" => Self::Exactness,
|
||||||
|
text => {
|
||||||
|
let caps = ASC_DESC_REGEX.captures(text).ok_or(())?;
|
||||||
|
let order = caps.get(1).unwrap().as_str();
|
||||||
|
let field_name = caps.get(2).unwrap().as_str();
|
||||||
|
match order {
|
||||||
|
"asc" => Self::Asc(field_name.to_string()),
|
||||||
|
"desc" => Self::Desc(field_name.to_string()),
|
||||||
|
_ => return Err(()),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// Any value that is present is considered Some value, including null.
|
// Any value that is present is considered Some value, including null.
|
||||||
fn deserialize_some<'de, T, D>(deserializer: D) -> StdResult<Option<T>, D::Error>
|
fn deserialize_some<'de, T, D>(deserializer: D) -> StdResult<Option<T>, D::Error>
|
||||||
where
|
where
|
||||||
|
@ -0,0 +1,24 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/v1/mod.rs
|
||||||
|
expression: dnd_spells.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"wordsPosition",
|
||||||
|
"exactness"
|
||||||
|
],
|
||||||
|
"distinctAttribute": null,
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"attributesForFaceting": []
|
||||||
|
}
|
@ -0,0 +1,24 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/v1/mod.rs
|
||||||
|
expression: dnd_spells.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"wordsPosition",
|
||||||
|
"exactness"
|
||||||
|
],
|
||||||
|
"distinctAttribute": null,
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"attributesForFaceting": []
|
||||||
|
}
|
@ -0,0 +1,38 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/v1/mod.rs
|
||||||
|
expression: products.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"wordsPosition",
|
||||||
|
"exactness"
|
||||||
|
],
|
||||||
|
"distinctAttribute": null,
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {
|
||||||
|
"android": [
|
||||||
|
"phone",
|
||||||
|
"smartphone"
|
||||||
|
],
|
||||||
|
"iphone": [
|
||||||
|
"phone",
|
||||||
|
"smartphone"
|
||||||
|
],
|
||||||
|
"phone": [
|
||||||
|
"android",
|
||||||
|
"iphone",
|
||||||
|
"smartphone"
|
||||||
|
]
|
||||||
|
},
|
||||||
|
"attributesForFaceting": []
|
||||||
|
}
|
@ -0,0 +1,28 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/v1/mod.rs
|
||||||
|
expression: movies.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"wordsPosition",
|
||||||
|
"exactness",
|
||||||
|
"asc(release_date)"
|
||||||
|
],
|
||||||
|
"distinctAttribute": null,
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"attributesForFaceting": [
|
||||||
|
"id",
|
||||||
|
"genres"
|
||||||
|
]
|
||||||
|
}
|
@ -0,0 +1,28 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/v1/mod.rs
|
||||||
|
expression: movies.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"wordsPosition",
|
||||||
|
"exactness",
|
||||||
|
"asc(release_date)"
|
||||||
|
],
|
||||||
|
"distinctAttribute": null,
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"attributesForFaceting": [
|
||||||
|
"id",
|
||||||
|
"genres"
|
||||||
|
]
|
||||||
|
}
|
@ -0,0 +1,28 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/v1/mod.rs
|
||||||
|
expression: movies.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"wordsPosition",
|
||||||
|
"exactness",
|
||||||
|
"asc(release_date)"
|
||||||
|
],
|
||||||
|
"distinctAttribute": null,
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"attributesForFaceting": [
|
||||||
|
"id",
|
||||||
|
"genres"
|
||||||
|
]
|
||||||
|
}
|
@ -0,0 +1,24 @@
|
|||||||
|
---
|
||||||
|
source: dump/src/reader/v1/mod.rs
|
||||||
|
expression: dnd_spells.settings().unwrap()
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"rankingRules": [
|
||||||
|
"typo",
|
||||||
|
"words",
|
||||||
|
"proximity",
|
||||||
|
"attribute",
|
||||||
|
"wordsPosition",
|
||||||
|
"exactness"
|
||||||
|
],
|
||||||
|
"distinctAttribute": null,
|
||||||
|
"searchableAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"displayedAttributes": [
|
||||||
|
"*"
|
||||||
|
],
|
||||||
|
"stopWords": [],
|
||||||
|
"synonyms": {},
|
||||||
|
"attributesForFaceting": []
|
||||||
|
}
|
@ -1,54 +1,8 @@
|
|||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use serde_json::Value;
|
|
||||||
use time::OffsetDateTime;
|
use time::OffsetDateTime;
|
||||||
|
|
||||||
use super::settings::SettingsUpdate;
|
use super::settings::SettingsUpdate;
|
||||||
|
|
||||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
|
||||||
pub struct Update {
|
|
||||||
data: UpdateData,
|
|
||||||
#[serde(with = "time::serde::rfc3339")]
|
|
||||||
enqueued_at: OffsetDateTime,
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
|
||||||
pub enum UpdateData {
|
|
||||||
ClearAll,
|
|
||||||
Customs(Vec<u8>),
|
|
||||||
// (primary key, documents)
|
|
||||||
DocumentsAddition {
|
|
||||||
primary_key: Option<String>,
|
|
||||||
documents: Vec<serde_json::Map<String, Value>>,
|
|
||||||
},
|
|
||||||
DocumentsPartial {
|
|
||||||
primary_key: Option<String>,
|
|
||||||
documents: Vec<serde_json::Map<String, Value>>,
|
|
||||||
},
|
|
||||||
DocumentsDeletion(Vec<String>),
|
|
||||||
Settings(Box<SettingsUpdate>),
|
|
||||||
}
|
|
||||||
|
|
||||||
impl UpdateData {
|
|
||||||
pub fn update_type(&self) -> UpdateType {
|
|
||||||
match self {
|
|
||||||
UpdateData::ClearAll => UpdateType::ClearAll,
|
|
||||||
UpdateData::Customs(_) => UpdateType::Customs,
|
|
||||||
UpdateData::DocumentsAddition { documents, .. } => UpdateType::DocumentsAddition {
|
|
||||||
number: documents.len(),
|
|
||||||
},
|
|
||||||
UpdateData::DocumentsPartial { documents, .. } => UpdateType::DocumentsPartial {
|
|
||||||
number: documents.len(),
|
|
||||||
},
|
|
||||||
UpdateData::DocumentsDeletion(deletion) => UpdateType::DocumentsDeletion {
|
|
||||||
number: deletion.len(),
|
|
||||||
},
|
|
||||||
UpdateData::Settings(update) => UpdateType::Settings {
|
|
||||||
settings: update.clone(),
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||||
#[serde(tag = "name")]
|
#[serde(tag = "name")]
|
||||||
pub enum UpdateType {
|
pub enum UpdateType {
|
||||||
|
@ -1,22 +0,0 @@
|
|||||||
use serde::Deserialize;
|
|
||||||
use time::OffsetDateTime;
|
|
||||||
|
|
||||||
#[derive(Debug, Deserialize)]
|
|
||||||
#[serde(rename_all = "camelCase")]
|
|
||||||
pub struct Index {
|
|
||||||
pub name: String,
|
|
||||||
pub uid: String,
|
|
||||||
#[serde(with = "time::serde::rfc3339")]
|
|
||||||
created_at: OffsetDateTime,
|
|
||||||
#[serde(with = "time::serde::rfc3339")]
|
|
||||||
updated_at: OffsetDateTime,
|
|
||||||
pub primary_key: Option<String>,
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug, Deserialize)]
|
|
||||||
#[serde(rename_all = "camelCase")]
|
|
||||||
pub struct Metadata {
|
|
||||||
indexes: Vec<Index>,
|
|
||||||
db_version: String,
|
|
||||||
dump_version: crate::Version,
|
|
||||||
}
|
|
@ -1,4 +1,5 @@
|
|||||||
use std::collections::{BTreeMap, BTreeSet};
|
use std::collections::{BTreeMap, BTreeSet};
|
||||||
|
use std::fmt::Display;
|
||||||
use std::marker::PhantomData;
|
use std::marker::PhantomData;
|
||||||
use std::str::FromStr;
|
use std::str::FromStr;
|
||||||
|
|
||||||
@ -174,3 +175,17 @@ impl FromStr for Criterion {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl Display for Criterion {
|
||||||
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
|
match self {
|
||||||
|
Criterion::Words => write!(f, "words"),
|
||||||
|
Criterion::Typo => write!(f, "typo"),
|
||||||
|
Criterion::Proximity => write!(f, "proximity"),
|
||||||
|
Criterion::Attribute => write!(f, "attribute"),
|
||||||
|
Criterion::Exactness => write!(f, "exactness"),
|
||||||
|
Criterion::Asc(field_name) => write!(f, "asc({})", field_name),
|
||||||
|
Criterion::Desc(field_name) => write!(f, "desc({})", field_name),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
BIN
dump/tests/assets/v1.dump
Normal file
BIN
dump/tests/assets/v1.dump
Normal file
Binary file not shown.
@ -9,19 +9,189 @@ use crate::common::{default_settings, GetAllDocumentsOptions, Server};
|
|||||||
// all the following test are ignored on windows. See #2364
|
// all the following test are ignored on windows. See #2364
|
||||||
#[actix_rt::test]
|
#[actix_rt::test]
|
||||||
#[cfg_attr(target_os = "windows", ignore)]
|
#[cfg_attr(target_os = "windows", ignore)]
|
||||||
async fn import_dump_v1() {
|
async fn import_dump_v1_movie_raw() {
|
||||||
|
let temp = tempfile::tempdir().unwrap();
|
||||||
|
let path = GetDump::MoviesRawV1.path();
|
||||||
|
let options = Opt { import_dump: Some(path), ..default_settings(temp.path()) };
|
||||||
|
let server = Server::new_with_options(options).await.unwrap();
|
||||||
|
|
||||||
|
let (indexes, code) = server.list_indexes(None, None).await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
|
||||||
|
assert_eq!(indexes["results"].as_array().unwrap().len(), 1);
|
||||||
|
assert_eq!(indexes["results"][0]["uid"], json!("indexUID"));
|
||||||
|
assert_eq!(indexes["results"][0]["primaryKey"], json!("id"));
|
||||||
|
|
||||||
|
let index = server.index("indexUID");
|
||||||
|
|
||||||
|
let (stats, code) = index.stats().await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
stats,
|
||||||
|
json!({ "numberOfDocuments": 53, "isIndexing": false, "fieldDistribution": {"genres": 53, "id": 53, "overview": 53, "poster": 53, "release_date": 53, "title": 53 }})
|
||||||
|
);
|
||||||
|
|
||||||
|
let (settings, code) = index.settings().await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
settings,
|
||||||
|
json!({"displayedAttributes": ["*"], "searchableAttributes": ["*"], "filterableAttributes": [], "sortableAttributes": [], "rankingRules": ["typo", "words", "proximity", "attribute", "exactness"], "stopWords": [], "synonyms": {}, "distinctAttribute": null, "typoTolerance": {"enabled": true, "minWordSizeForTypos": {"oneTypo": 5, "twoTypos": 9}, "disableOnWords": [], "disableOnAttributes": [] }, "faceting": { "maxValuesPerFacet": 100 }, "pagination": { "maxTotalHits": 1000 } })
|
||||||
|
);
|
||||||
|
|
||||||
|
let (tasks, code) = index.list_tasks().await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
tasks,
|
||||||
|
json!({ "results": [{"uid": 0, "indexUid": "indexUID", "status": "succeeded", "type": "documentAdditionOrUpdate", "canceledBy": null, "details": { "receivedDocuments": 0, "indexedDocuments": 31968 }, "error": null, "duration": "PT9.317060500S", "enqueuedAt": "2021-09-08T09:08:45.153219Z", "startedAt": "2021-09-08T09:08:45.3961665Z", "finishedAt": "2021-09-08T09:08:54.713227Z" }], "limit": 20, "from": 0, "next": null })
|
||||||
|
);
|
||||||
|
|
||||||
|
// finally we're just going to check that we can still get a few documents by id
|
||||||
|
let (document, code) = index.get_document(100, None).await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
document,
|
||||||
|
json!({"id": 100, "title": "Lock, Stock and Two Smoking Barrels", "overview": "A card shark and his unwillingly-enlisted friends need to make a lot of cash quick after losing a sketchy poker match. To do this they decide to pull a heist on a small-time gang who happen to be operating out of the flat next door.", "genres": ["Comedy", "Crime"], "poster": "https://image.tmdb.org/t/p/w500/8kSerJrhrJWKLk1LViesGcnrUPE.jpg", "release_date": 889056000})
|
||||||
|
);
|
||||||
|
|
||||||
|
let (document, code) = index.get_document(500, None).await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
document,
|
||||||
|
json!({"id": 500, "title": "Reservoir Dogs", "overview": "A botched robbery indicates a police informant, and the pressure mounts in the aftermath at a warehouse. Crime begets violence as the survivors -- veteran Mr. White, newcomer Mr. Orange, psychopathic parolee Mr. Blonde, bickering weasel Mr. Pink and Nice Guy Eddie -- unravel.", "genres": ["Crime", "Thriller"], "poster": "https://image.tmdb.org/t/p/w500/AjTtJNumZyUDz33VtMlF1K8JPsE.jpg", "release_date": 715392000})
|
||||||
|
);
|
||||||
|
|
||||||
|
let (document, code) = index.get_document(10006, None).await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
document,
|
||||||
|
json!({"id": 10006, "title": "Wild Seven", "overview": "In this darkly karmic vision of Arizona, a man who breathes nothing but ill will begins a noxious domino effect as quickly as an uncontrollable virus kills. As he exits Arizona State Penn after twenty-one long years, Wilson has only one thing on the brain, leveling the score with career criminal, Mackey Willis.", "genres": ["Action", "Crime", "Drama"], "poster": "https://image.tmdb.org/t/p/w500/y114dTPoqn8k2Txps4P2tI95YCS.jpg", "release_date": 1136073600})
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[actix_rt::test]
|
||||||
|
#[cfg_attr(target_os = "windows", ignore)]
|
||||||
|
async fn import_dump_v1_movie_with_settings() {
|
||||||
let temp = tempfile::tempdir().unwrap();
|
let temp = tempfile::tempdir().unwrap();
|
||||||
|
|
||||||
for path in [
|
let path = GetDump::MoviesWithSettingsV1.path();
|
||||||
GetDump::MoviesRawV1.path(),
|
|
||||||
GetDump::MoviesWithSettingsV1.path(),
|
|
||||||
GetDump::RubyGemsWithSettingsV1.path(),
|
|
||||||
] {
|
|
||||||
let options = Opt { import_dump: Some(path), ..default_settings(temp.path()) };
|
|
||||||
let error = Server::new_with_options(options).await.map(drop).unwrap_err();
|
|
||||||
|
|
||||||
assert_eq!(error.to_string(), "The version 1 of the dumps is not supported anymore. You can re-export your dump from a version between 0.21 and 0.24, or start fresh from a version 0.25 onwards.");
|
let options = Opt { import_dump: Some(path), ..default_settings(temp.path()) };
|
||||||
}
|
let server = Server::new_with_options(options).await.unwrap();
|
||||||
|
|
||||||
|
let (indexes, code) = server.list_indexes(None, None).await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
|
||||||
|
assert_eq!(indexes["results"].as_array().unwrap().len(), 1);
|
||||||
|
assert_eq!(indexes["results"][0]["uid"], json!("indexUID"));
|
||||||
|
assert_eq!(indexes["results"][0]["primaryKey"], json!("id"));
|
||||||
|
|
||||||
|
let index = server.index("indexUID");
|
||||||
|
|
||||||
|
let (stats, code) = index.stats().await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
stats,
|
||||||
|
json!({ "numberOfDocuments": 53, "isIndexing": false, "fieldDistribution": {"genres": 53, "id": 53, "overview": 53, "poster": 53, "release_date": 53, "title": 53 }})
|
||||||
|
);
|
||||||
|
|
||||||
|
let (settings, code) = index.settings().await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
settings,
|
||||||
|
json!({ "displayedAttributes": ["genres", "id", "overview", "poster", "release_date", "title"], "searchableAttributes": ["title", "overview"], "filterableAttributes": ["genres"], "sortableAttributes": [], "rankingRules": ["typo", "words", "proximity", "attribute", "exactness"], "stopWords": ["of", "the"], "synonyms": {}, "distinctAttribute": null, "typoTolerance": {"enabled": true, "minWordSizeForTypos": { "oneTypo": 5, "twoTypos": 9 }, "disableOnWords": [], "disableOnAttributes": [] }, "faceting": { "maxValuesPerFacet": 100 }, "pagination": { "maxTotalHits": 1000 } })
|
||||||
|
);
|
||||||
|
|
||||||
|
let (tasks, code) = index.list_tasks().await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
tasks,
|
||||||
|
json!({ "results": [{ "uid": 1, "indexUid": "indexUID", "status": "succeeded", "type": "settingsUpdate", "canceledBy": null, "details": { "displayedAttributes": ["genres", "id", "overview", "poster", "release_date", "title"], "searchableAttributes": ["title", "overview"], "filterableAttributes": ["genres"], "stopWords": ["of", "the"] }, "error": null, "duration": "PT7.288826907S", "enqueuedAt": "2021-09-08T09:34:40.882977Z", "startedAt": "2021-09-08T09:34:40.883073093Z", "finishedAt": "2021-09-08T09:34:48.1719Z"}, { "uid": 0, "indexUid": "indexUID", "status": "succeeded", "type": "documentAdditionOrUpdate", "canceledBy": null, "details": { "receivedDocuments": 0, "indexedDocuments": 31968 }, "error": null, "duration": "PT9.090735774S", "enqueuedAt": "2021-09-08T09:34:16.036101Z", "startedAt": "2021-09-08T09:34:16.261191226Z", "finishedAt": "2021-09-08T09:34:25.351927Z" }], "limit": 20, "from": 1, "next": null })
|
||||||
|
);
|
||||||
|
|
||||||
|
// finally we're just going to check that we can still get a few documents by id
|
||||||
|
let (document, code) = index.get_document(100, None).await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
document,
|
||||||
|
json!({ "id": 100, "title": "Lock, Stock and Two Smoking Barrels", "genres": ["Comedy", "Crime"], "overview": "A card shark and his unwillingly-enlisted friends need to make a lot of cash quick after losing a sketchy poker match. To do this they decide to pull a heist on a small-time gang who happen to be operating out of the flat next door.", "poster": "https://image.tmdb.org/t/p/w500/8kSerJrhrJWKLk1LViesGcnrUPE.jpg", "release_date": 889056000 })
|
||||||
|
);
|
||||||
|
|
||||||
|
let (document, code) = index.get_document(500, None).await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
document,
|
||||||
|
json!({ "id": 500, "title": "Reservoir Dogs", "genres": ["Crime", "Thriller"], "overview": "A botched robbery indicates a police informant, and the pressure mounts in the aftermath at a warehouse. Crime begets violence as the survivors -- veteran Mr. White, newcomer Mr. Orange, psychopathic parolee Mr. Blonde, bickering weasel Mr. Pink and Nice Guy Eddie -- unravel.", "poster": "https://image.tmdb.org/t/p/w500/AjTtJNumZyUDz33VtMlF1K8JPsE.jpg", "release_date": 715392000})
|
||||||
|
);
|
||||||
|
|
||||||
|
let (document, code) = index.get_document(10006, None).await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
document,
|
||||||
|
json!({ "id": 10006, "title": "Wild Seven", "genres": ["Action", "Crime", "Drama"], "overview": "In this darkly karmic vision of Arizona, a man who breathes nothing but ill will begins a noxious domino effect as quickly as an uncontrollable virus kills. As he exits Arizona State Penn after twenty-one long years, Wilson has only one thing on the brain, leveling the score with career criminal, Mackey Willis.", "poster": "https://image.tmdb.org/t/p/w500/y114dTPoqn8k2Txps4P2tI95YCS.jpg", "release_date": 1136073600})
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[actix_rt::test]
|
||||||
|
#[cfg_attr(target_os = "windows", ignore)]
|
||||||
|
async fn import_dump_v1_rubygems_with_settings() {
|
||||||
|
let temp = tempfile::tempdir().unwrap();
|
||||||
|
|
||||||
|
let path = GetDump::RubyGemsWithSettingsV1.path();
|
||||||
|
|
||||||
|
let options = Opt { import_dump: Some(path), ..default_settings(temp.path()) };
|
||||||
|
let server = Server::new_with_options(options).await.unwrap();
|
||||||
|
|
||||||
|
let (indexes, code) = server.list_indexes(None, None).await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
|
||||||
|
assert_eq!(indexes["results"].as_array().unwrap().len(), 1);
|
||||||
|
assert_eq!(indexes["results"][0]["uid"], json!("rubygems"));
|
||||||
|
assert_eq!(indexes["results"][0]["primaryKey"], json!("id"));
|
||||||
|
|
||||||
|
let index = server.index("rubygems");
|
||||||
|
|
||||||
|
let (stats, code) = index.stats().await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
stats,
|
||||||
|
json!({ "numberOfDocuments": 53, "isIndexing": false, "fieldDistribution": {"description": 53, "id": 53, "name": 53, "summary": 53, "total_downloads": 53, "version": 53 }})
|
||||||
|
);
|
||||||
|
|
||||||
|
let (settings, code) = index.settings().await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
settings,
|
||||||
|
json!({"displayedAttributes": ["description", "id", "name", "summary", "total_downloads", "version"], "searchableAttributes": ["name", "summary"], "filterableAttributes": ["version"], "sortableAttributes": [], "rankingRules": ["typo", "words", "fame:desc", "proximity", "attribute", "exactness", "total_downloads:desc"], "stopWords": [], "synonyms": {}, "distinctAttribute": null, "typoTolerance": {"enabled": true, "minWordSizeForTypos": {"oneTypo": 5, "twoTypos": 9}, "disableOnWords": [], "disableOnAttributes": [] }, "faceting": { "maxValuesPerFacet": 100 }, "pagination": { "maxTotalHits": 1000 }})
|
||||||
|
);
|
||||||
|
|
||||||
|
let (tasks, code) = index.list_tasks().await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
tasks["results"][0],
|
||||||
|
json!({"uid": 92, "indexUid": "rubygems", "status": "succeeded", "type": "documentAdditionOrUpdate", "canceledBy": null, "details": {"receivedDocuments": 0, "indexedDocuments": 1042}, "error": null, "duration": "PT1.487793839S", "enqueuedAt": "2021-09-08T09:27:01.465296Z", "startedAt": "2021-09-08T09:28:44.882177161Z", "finishedAt": "2021-09-08T09:28:46.369971Z"})
|
||||||
|
);
|
||||||
|
|
||||||
|
// finally we're just going to check that we can still get a few documents by id
|
||||||
|
let (document, code) = index.get_document(188040, None).await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
document,
|
||||||
|
json!({ "name": "meilisearch", "summary": "An easy-to-use ruby client for Meilisearch API", "description": "An easy-to-use ruby client for Meilisearch API. See https://github.com/meilisearch/MeiliSearch", "id": "188040", "version": "0.15.2", "total_downloads": "7465"})
|
||||||
|
);
|
||||||
|
|
||||||
|
let (document, code) = index.get_document(191940, None).await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
document,
|
||||||
|
json!({ "name": "doggo", "summary": "RSpec 3 formatter - documentation, with progress indication", "description": "Similar to \"rspec -f d\", but also indicates progress by showing the current test number and total test count on each line.", "id": "191940", "version": "1.1.0", "total_downloads": "9394"})
|
||||||
|
);
|
||||||
|
|
||||||
|
let (document, code) = index.get_document(159227, None).await;
|
||||||
|
assert_eq!(code, 200);
|
||||||
|
assert_eq!(
|
||||||
|
document,
|
||||||
|
json!({ "name": "vortex-of-agony", "summary": "You dont need to use nodejs or go, just install this plugin. It will crash your application at random", "description": "You dont need to use nodejs or go, just install this plugin. It will crash your application at random", "id": "159227", "version": "0.1.0", "total_downloads": "1007"})
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[actix_rt::test]
|
#[actix_rt::test]
|
||||||
|
Loading…
Reference in New Issue
Block a user