mirror of
https://github.com/meilisearch/meilisearch.git
synced 2024-11-23 18:45:06 +08:00
first version working with index operation
This commit is contained in:
parent
f1aa22567f
commit
301907869d
@ -103,3 +103,46 @@ pub enum IndexOperation {
|
|||||||
settings_tasks: Vec<TaskId>,
|
settings_tasks: Vec<TaskId>,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl Batch {
|
||||||
|
pub fn ids(&self) -> impl Iterator<Item = TaskId> {
|
||||||
|
type Ret = Box<dyn Iterator<Item = TaskId>>;
|
||||||
|
|
||||||
|
match self {
|
||||||
|
Batch::TaskCancelation { task, .. } => Box::new(std::iter::once(*task)) as Ret,
|
||||||
|
Batch::TaskDeletion(task) => Box::new(std::iter::once(*task)) as Ret,
|
||||||
|
Batch::SnapshotCreation(tasks) => Box::new(tasks.clone().into_iter()) as Ret,
|
||||||
|
Batch::Dump(task) => Box::new(std::iter::once(*task)) as Ret,
|
||||||
|
Batch::IndexOperation { op, .. } => match op {
|
||||||
|
IndexOperation::DocumentOperation { tasks, .. } => {
|
||||||
|
Box::new(tasks.clone().into_iter()) as Ret
|
||||||
|
}
|
||||||
|
IndexOperation::DocumentDeletion { tasks, .. } => {
|
||||||
|
Box::new(tasks.clone().into_iter()) as Ret
|
||||||
|
}
|
||||||
|
IndexOperation::DocumentClear { tasks, .. } => {
|
||||||
|
Box::new(tasks.clone().into_iter()) as Ret
|
||||||
|
}
|
||||||
|
IndexOperation::Settings { tasks, .. } => {
|
||||||
|
Box::new(tasks.clone().into_iter()) as Ret
|
||||||
|
}
|
||||||
|
IndexOperation::DocumentClearAndSetting {
|
||||||
|
cleared_tasks, settings_tasks, ..
|
||||||
|
} => {
|
||||||
|
Box::new(cleared_tasks.clone().into_iter().chain(settings_tasks.clone())) as Ret
|
||||||
|
}
|
||||||
|
IndexOperation::SettingsAndDocumentOperation {
|
||||||
|
document_import_tasks,
|
||||||
|
settings_tasks,
|
||||||
|
..
|
||||||
|
} => Box::new(
|
||||||
|
document_import_tasks.clone().into_iter().chain(settings_tasks.clone()),
|
||||||
|
) as Ret,
|
||||||
|
},
|
||||||
|
Batch::IndexCreation { task, .. } => Box::new(std::iter::once(*task)) as Ret,
|
||||||
|
Batch::IndexUpdate { task, .. } => Box::new(std::iter::once(*task)) as Ret,
|
||||||
|
Batch::IndexDeletion { tasks, .. } => Box::new(tasks.clone().into_iter()) as Ret,
|
||||||
|
Batch::IndexSwap { task } => Box::new(std::iter::once(*task)) as Ret,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
use std::net::ToSocketAddrs;
|
use std::net::ToSocketAddrs;
|
||||||
use std::sync::atomic::AtomicUsize;
|
use std::sync::atomic::{AtomicUsize, Ordering};
|
||||||
use std::sync::{atomic, Arc, Mutex};
|
use std::sync::{atomic, Arc, Mutex, RwLock};
|
||||||
|
|
||||||
use bus::{Bus, BusReader};
|
use bus::{Bus, BusReader};
|
||||||
use crossbeam::channel::{unbounded, Receiver, Sender};
|
use crossbeam::channel::{unbounded, Receiver, Sender};
|
||||||
@ -11,14 +11,14 @@ use meilisearch_types::tasks::Task;
|
|||||||
use crate::batch::Batch;
|
use crate::batch::Batch;
|
||||||
use crate::{Consistency, FollowerMsg, LeaderMsg};
|
use crate::{Consistency, FollowerMsg, LeaderMsg};
|
||||||
|
|
||||||
#[derive(Clone)]
|
#[derive(Clone, Debug)]
|
||||||
pub struct Leader {
|
pub struct Leader {
|
||||||
task_ready_to_commit: Receiver<u32>,
|
task_ready_to_commit: Receiver<u32>,
|
||||||
broadcast_to_follower: Sender<LeaderMsg>,
|
broadcast_to_follower: Sender<LeaderMsg>,
|
||||||
|
|
||||||
cluster_size: Arc<AtomicUsize>,
|
cluster_size: Arc<AtomicUsize>,
|
||||||
|
|
||||||
batch_id: u32,
|
batch_id: Arc<RwLock<u32>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Leader {
|
impl Leader {
|
||||||
@ -36,7 +36,7 @@ impl Leader {
|
|||||||
task_ready_to_commit: task_finished_receiver,
|
task_ready_to_commit: task_finished_receiver,
|
||||||
broadcast_to_follower: process_batch_sender,
|
broadcast_to_follower: process_batch_sender,
|
||||||
cluster_size,
|
cluster_size,
|
||||||
batch_id: 0,
|
batch_id: Arc::default(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -113,46 +113,62 @@ impl Leader {
|
|||||||
info!("A follower left the cluster. {} members.", size);
|
info!("A follower left the cluster. {} members.", size);
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn starts_batch(&mut self, batch: Batch) {
|
pub fn starts_batch(&self, batch: Batch) {
|
||||||
|
let mut batch_id = self.batch_id.write().unwrap();
|
||||||
|
|
||||||
assert!(
|
assert!(
|
||||||
self.batch_id % 2 == 0,
|
*batch_id % 2 == 0,
|
||||||
"Tried to start processing a batch before commiting the previous one"
|
"Tried to start processing a batch before commiting the previous one"
|
||||||
);
|
);
|
||||||
self.batch_id += 1;
|
info!("Send the batch to process to the followers");
|
||||||
|
*batch_id += 1;
|
||||||
|
|
||||||
self.broadcast_to_follower
|
self.broadcast_to_follower
|
||||||
.send(LeaderMsg::StartBatch { id: self.batch_id, batch })
|
.send(LeaderMsg::StartBatch { id: *batch_id, batch })
|
||||||
.expect("Can't reach the cluster");
|
.expect("Can't reach the cluster");
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn commit(&mut self, consistency_level: Consistency) {
|
pub fn commit(&self, consistency_level: Consistency) {
|
||||||
|
info!("Wait until enough followers are ready to commit a batch");
|
||||||
|
|
||||||
|
let mut batch_id = self.batch_id.write().unwrap();
|
||||||
|
|
||||||
// if zero nodes needs to be sync we can commit right away and early exit
|
// if zero nodes needs to be sync we can commit right away and early exit
|
||||||
if consistency_level != Consistency::Zero {
|
if consistency_level != Consistency::One {
|
||||||
// else, we wait till enough nodes are ready to commit
|
// else, we wait till enough nodes are ready to commit
|
||||||
for (ready_to_commit, _id) in self
|
for ready_to_commit in self
|
||||||
.task_ready_to_commit
|
.task_ready_to_commit
|
||||||
.iter()
|
.iter()
|
||||||
// we need to filter out the messages from the old batches
|
// we need to filter out the messages from the old batches
|
||||||
.filter(|id| *id == self.batch_id)
|
.filter(|id| *id == *batch_id)
|
||||||
.enumerate()
|
.enumerate()
|
||||||
|
// we do a +2 because enumerate starts at 1 and we must includes ourselves in the count
|
||||||
|
.map(|(id, _)| id + 2)
|
||||||
{
|
{
|
||||||
|
// TODO: if the last node dies we're stuck on the iterator
|
||||||
|
|
||||||
|
// we need to reload the cluster size everytime in case a node dies
|
||||||
let cluster_size = self.cluster_size.load(atomic::Ordering::Relaxed);
|
let cluster_size = self.cluster_size.load(atomic::Ordering::Relaxed);
|
||||||
|
|
||||||
|
info!("{ready_to_commit} nodes are ready to commit for a cluster size of {cluster_size}");
|
||||||
match consistency_level {
|
match consistency_level {
|
||||||
Consistency::One if ready_to_commit >= 1 => break,
|
Consistency::Two if ready_to_commit >= 1 => break,
|
||||||
Consistency::Two if ready_to_commit >= 2 => break,
|
|
||||||
Consistency::Quorum if ready_to_commit >= (cluster_size / 2) => break,
|
Consistency::Quorum if ready_to_commit >= (cluster_size / 2) => break,
|
||||||
|
Consistency::All if ready_to_commit == cluster_size => break,
|
||||||
_ => (),
|
_ => (),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
self.broadcast_to_follower.send(LeaderMsg::Commit(self.batch_id)).unwrap();
|
info!("Tells all the follower to commit");
|
||||||
|
|
||||||
self.batch_id += 1;
|
self.broadcast_to_follower.send(LeaderMsg::Commit(*batch_id)).unwrap();
|
||||||
|
|
||||||
|
*batch_id += 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn register_new_task(&mut self, task: Task, update_file: Option<Vec<u8>>) {
|
pub fn register_new_task(&self, task: Task, update_file: Option<Vec<u8>>) {
|
||||||
|
info!("Tells all the follower to register a new task");
|
||||||
self.broadcast_to_follower
|
self.broadcast_to_follower
|
||||||
.send(LeaderMsg::RegisterNewTask { task, update_file })
|
.send(LeaderMsg::RegisterNewTask { task, update_file })
|
||||||
.expect("Main thread is dead");
|
.expect("Main thread is dead");
|
||||||
|
@ -1,4 +1,5 @@
|
|||||||
use std::net::ToSocketAddrs;
|
use std::net::ToSocketAddrs;
|
||||||
|
use std::sync::{Arc, RwLock};
|
||||||
|
|
||||||
use batch::Batch;
|
use batch::Batch;
|
||||||
use crossbeam::channel::{unbounded, Receiver, Sender};
|
use crossbeam::channel::{unbounded, Receiver, Sender};
|
||||||
@ -39,7 +40,6 @@ pub enum FollowerMsg {
|
|||||||
|
|
||||||
#[derive(Debug, Clone, Copy, PartialEq, Eq)]
|
#[derive(Debug, Clone, Copy, PartialEq, Eq)]
|
||||||
pub enum Consistency {
|
pub enum Consistency {
|
||||||
Zero,
|
|
||||||
One,
|
One,
|
||||||
Two,
|
Two,
|
||||||
Quorum,
|
Quorum,
|
||||||
@ -54,7 +54,7 @@ pub struct Follower {
|
|||||||
must_commit: Receiver<u32>,
|
must_commit: Receiver<u32>,
|
||||||
register_new_task: Receiver<(Task, Option<Vec<u8>>)>,
|
register_new_task: Receiver<(Task, Option<Vec<u8>>)>,
|
||||||
|
|
||||||
batch_id: u32,
|
batch_id: Arc<RwLock<u32>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Follower {
|
impl Follower {
|
||||||
@ -76,7 +76,7 @@ impl Follower {
|
|||||||
get_batch: get_batch_receiver,
|
get_batch: get_batch_receiver,
|
||||||
must_commit: must_commit_receiver,
|
must_commit: must_commit_receiver,
|
||||||
register_new_task: register_task_receiver,
|
register_new_task: register_task_receiver,
|
||||||
batch_id: 0,
|
batch_id: Arc::default(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -106,31 +106,33 @@ impl Follower {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_new_batch(&mut self) -> Batch {
|
pub fn get_new_batch(&self) -> Batch {
|
||||||
info!("Get new batch called");
|
info!("Get new batch called");
|
||||||
let (id, batch) = self.get_batch.recv().expect("Lost connection to the leader");
|
let (id, batch) = self.get_batch.recv().expect("Lost connection to the leader");
|
||||||
info!("Got a new batch");
|
info!("Got a new batch");
|
||||||
self.batch_id = id;
|
*self.batch_id.write().unwrap() = id;
|
||||||
batch
|
batch
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn ready_to_commit(&mut self) {
|
pub fn ready_to_commit(&self) {
|
||||||
info!("I'm ready to commit");
|
info!("I'm ready to commit");
|
||||||
self.sender.send(FollowerMsg::ReadyToCommit(self.batch_id)).unwrap();
|
let batch_id = self.batch_id.read().unwrap();
|
||||||
|
|
||||||
|
self.sender.send(FollowerMsg::ReadyToCommit(*batch_id)).unwrap();
|
||||||
|
|
||||||
loop {
|
loop {
|
||||||
let id = self.must_commit.recv().expect("Lost connection to the leader");
|
let id = self.must_commit.recv().expect("Lost connection to the leader");
|
||||||
#[allow(clippy::comparison_chain)]
|
#[allow(clippy::comparison_chain)]
|
||||||
if id == self.batch_id {
|
if id == *batch_id {
|
||||||
break;
|
break;
|
||||||
} else if id > self.batch_id {
|
} else if id > *batch_id {
|
||||||
panic!("We missed a batch");
|
panic!("We missed a batch");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
info!("I got the right to commit");
|
info!("I got the right to commit");
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_new_task(&mut self) -> (Task, Option<Vec<u8>>) {
|
pub fn get_new_task(&self) -> (Task, Option<Vec<u8>>) {
|
||||||
self.register_new_task.recv().unwrap()
|
self.register_new_task.recv().unwrap()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -14,6 +14,7 @@ license.workspace = true
|
|||||||
anyhow = "1.0.64"
|
anyhow = "1.0.64"
|
||||||
bincode = "1.3.3"
|
bincode = "1.3.3"
|
||||||
cluster = { path = "../cluster" }
|
cluster = { path = "../cluster" }
|
||||||
|
crossbeam = "0.8.2"
|
||||||
csv = "1.1.6"
|
csv = "1.1.6"
|
||||||
derive_builder = "0.11.2"
|
derive_builder = "0.11.2"
|
||||||
dump = { path = "../dump" }
|
dump = { path = "../dump" }
|
||||||
|
@ -23,6 +23,7 @@ use std::fs::{self, File};
|
|||||||
use std::io::BufWriter;
|
use std::io::BufWriter;
|
||||||
|
|
||||||
use cluster::Consistency;
|
use cluster::Consistency;
|
||||||
|
use crossbeam::utils::Backoff;
|
||||||
use dump::IndexMetadata;
|
use dump::IndexMetadata;
|
||||||
use log::{debug, error, info};
|
use log::{debug, error, info};
|
||||||
use meilisearch_types::heed::{RoTxn, RwTxn};
|
use meilisearch_types::heed::{RoTxn, RwTxn};
|
||||||
@ -588,12 +589,8 @@ impl IndexScheduler {
|
|||||||
}
|
}
|
||||||
|
|
||||||
match &self.cluster {
|
match &self.cluster {
|
||||||
Some(Cluster::Leader(leader)) => {
|
Some(Cluster::Leader(leader)) => leader.commit(Consistency::All),
|
||||||
leader.write().unwrap().commit(Consistency::All)
|
Some(Cluster::Follower(follower)) => follower.ready_to_commit(),
|
||||||
}
|
|
||||||
Some(Cluster::Follower(follower)) => {
|
|
||||||
follower.write().unwrap().ready_to_commit()
|
|
||||||
}
|
|
||||||
None => (),
|
None => (),
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -642,12 +639,8 @@ impl IndexScheduler {
|
|||||||
}
|
}
|
||||||
|
|
||||||
match &self.cluster {
|
match &self.cluster {
|
||||||
Some(Cluster::Leader(leader)) => {
|
Some(Cluster::Leader(leader)) => leader.commit(Consistency::All),
|
||||||
leader.write().unwrap().commit(Consistency::All)
|
Some(Cluster::Follower(follower)) => follower.ready_to_commit(),
|
||||||
}
|
|
||||||
Some(Cluster::Follower(follower)) => {
|
|
||||||
follower.write().unwrap().ready_to_commit()
|
|
||||||
}
|
|
||||||
None => (),
|
None => (),
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -864,12 +857,8 @@ impl IndexScheduler {
|
|||||||
let tasks = self.apply_index_operation(&mut index_wtxn, &index, op)?;
|
let tasks = self.apply_index_operation(&mut index_wtxn, &index, op)?;
|
||||||
|
|
||||||
match &self.cluster {
|
match &self.cluster {
|
||||||
Some(Cluster::Leader(leader)) => {
|
Some(Cluster::Leader(leader)) => leader.commit(Consistency::All),
|
||||||
leader.write().unwrap().commit(Consistency::All)
|
Some(Cluster::Follower(follower)) => follower.ready_to_commit(),
|
||||||
}
|
|
||||||
Some(Cluster::Follower(follower)) => {
|
|
||||||
follower.write().unwrap().ready_to_commit()
|
|
||||||
}
|
|
||||||
None => (),
|
None => (),
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -973,12 +962,8 @@ impl IndexScheduler {
|
|||||||
}
|
}
|
||||||
|
|
||||||
match &self.cluster {
|
match &self.cluster {
|
||||||
Some(Cluster::Leader(leader)) => {
|
Some(Cluster::Leader(leader)) => leader.commit(Consistency::All),
|
||||||
leader.write().unwrap().commit(Consistency::All)
|
Some(Cluster::Follower(follower)) => follower.ready_to_commit(),
|
||||||
}
|
|
||||||
Some(Cluster::Follower(follower)) => {
|
|
||||||
follower.write().unwrap().ready_to_commit()
|
|
||||||
}
|
|
||||||
None => (),
|
None => (),
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1422,51 +1407,69 @@ impl IndexScheduler {
|
|||||||
|
|
||||||
pub(crate) fn get_batch_from_cluster_batch(
|
pub(crate) fn get_batch_from_cluster_batch(
|
||||||
&self,
|
&self,
|
||||||
rtxn: &RoTxn,
|
|
||||||
batch: cluster::batch::Batch,
|
batch: cluster::batch::Batch,
|
||||||
) -> Result<Batch> {
|
) -> Result<Batch> {
|
||||||
use cluster::batch::Batch as CBatch;
|
use cluster::batch::Batch as CBatch;
|
||||||
|
|
||||||
|
let mut rtxn = self.env.read_txn().map_err(Error::HeedTransaction)?;
|
||||||
|
|
||||||
|
for id in batch.ids() {
|
||||||
|
let backoff = Backoff::new();
|
||||||
|
let id = BEU32::new(id);
|
||||||
|
|
||||||
|
loop {
|
||||||
|
if self.all_tasks.get(&rtxn, &id)?.is_some() {
|
||||||
|
info!("Found the task_id");
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
info!("The task is not present in the task queue, we wait");
|
||||||
|
// we need to drop the txn to make a write visible
|
||||||
|
drop(rtxn);
|
||||||
|
backoff.spin();
|
||||||
|
rtxn = self.env.read_txn().map_err(Error::HeedTransaction)?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
Ok(match batch {
|
Ok(match batch {
|
||||||
CBatch::TaskCancelation { task, previous_started_at, previous_processing_tasks } => {
|
CBatch::TaskCancelation { task, previous_started_at, previous_processing_tasks } => {
|
||||||
Batch::TaskCancelation {
|
Batch::TaskCancelation {
|
||||||
task: self.get_existing_tasks(rtxn, Some(task))?[0].clone(),
|
task: self.get_existing_tasks(&rtxn, Some(task))?[0].clone(),
|
||||||
previous_started_at,
|
previous_started_at,
|
||||||
previous_processing_tasks,
|
previous_processing_tasks,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
CBatch::TaskDeletion(task) => {
|
CBatch::TaskDeletion(task) => {
|
||||||
Batch::TaskDeletion(self.get_existing_tasks(rtxn, Some(task))?[0].clone())
|
Batch::TaskDeletion(self.get_existing_tasks(&rtxn, Some(task))?[0].clone())
|
||||||
}
|
}
|
||||||
CBatch::SnapshotCreation(tasks) => {
|
CBatch::SnapshotCreation(tasks) => {
|
||||||
Batch::SnapshotCreation(self.get_existing_tasks(rtxn, tasks)?)
|
Batch::SnapshotCreation(self.get_existing_tasks(&rtxn, tasks)?)
|
||||||
}
|
}
|
||||||
CBatch::Dump(task) => {
|
CBatch::Dump(task) => {
|
||||||
Batch::Dump(self.get_existing_tasks(rtxn, Some(task))?[0].clone())
|
Batch::Dump(self.get_existing_tasks(&rtxn, Some(task))?[0].clone())
|
||||||
}
|
}
|
||||||
CBatch::IndexOperation { op, must_create_index } => Batch::IndexOperation {
|
CBatch::IndexOperation { op, must_create_index } => Batch::IndexOperation {
|
||||||
op: self.get_index_op_from_cluster_index_op(rtxn, op)?,
|
op: self.get_index_op_from_cluster_index_op(&rtxn, op)?,
|
||||||
must_create_index,
|
must_create_index,
|
||||||
},
|
},
|
||||||
CBatch::IndexCreation { index_uid, primary_key, task } => Batch::IndexCreation {
|
CBatch::IndexCreation { index_uid, primary_key, task } => Batch::IndexCreation {
|
||||||
index_uid,
|
index_uid,
|
||||||
primary_key,
|
primary_key,
|
||||||
task: self.get_existing_tasks(rtxn, Some(task))?[0].clone(),
|
task: self.get_existing_tasks(&rtxn, Some(task))?[0].clone(),
|
||||||
},
|
},
|
||||||
CBatch::IndexUpdate { index_uid, primary_key, task } => Batch::IndexUpdate {
|
CBatch::IndexUpdate { index_uid, primary_key, task } => Batch::IndexUpdate {
|
||||||
index_uid,
|
index_uid,
|
||||||
primary_key,
|
primary_key,
|
||||||
task: self.get_existing_tasks(rtxn, Some(task))?[0].clone(),
|
task: self.get_existing_tasks(&rtxn, Some(task))?[0].clone(),
|
||||||
},
|
},
|
||||||
CBatch::IndexDeletion { index_uid, tasks, index_has_been_created } => {
|
CBatch::IndexDeletion { index_uid, tasks, index_has_been_created } => {
|
||||||
Batch::IndexDeletion {
|
Batch::IndexDeletion {
|
||||||
index_uid,
|
index_uid,
|
||||||
tasks: self.get_existing_tasks(rtxn, tasks)?,
|
tasks: self.get_existing_tasks(&rtxn, tasks)?,
|
||||||
index_has_been_created,
|
index_has_been_created,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
CBatch::IndexSwap { task } => {
|
CBatch::IndexSwap { task } => {
|
||||||
Batch::IndexSwap { task: self.get_existing_tasks(rtxn, Some(task))?[0].clone() }
|
Batch::IndexSwap { task: self.get_existing_tasks(&rtxn, Some(task))?[0].clone() }
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
@ -349,8 +349,8 @@ impl std::str::FromStr for ClusterMode {
|
|||||||
|
|
||||||
#[derive(Clone)]
|
#[derive(Clone)]
|
||||||
pub enum Cluster {
|
pub enum Cluster {
|
||||||
Leader(Arc<RwLock<Leader>>),
|
Leader(Leader),
|
||||||
Follower(Arc<RwLock<Follower>>),
|
Follower(Follower),
|
||||||
}
|
}
|
||||||
|
|
||||||
impl IndexScheduler {
|
impl IndexScheduler {
|
||||||
@ -375,7 +375,7 @@ impl IndexScheduler {
|
|||||||
dumps_path: self.dumps_path.clone(),
|
dumps_path: self.dumps_path.clone(),
|
||||||
auth_path: self.auth_path.clone(),
|
auth_path: self.auth_path.clone(),
|
||||||
version_file_path: self.version_file_path.clone(),
|
version_file_path: self.version_file_path.clone(),
|
||||||
cluster: None,
|
cluster: self.cluster.clone(),
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
test_breakpoint_sdr: self.test_breakpoint_sdr.clone(),
|
test_breakpoint_sdr: self.test_breakpoint_sdr.clone(),
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
@ -544,11 +544,22 @@ impl IndexScheduler {
|
|||||||
fn run(&self) {
|
fn run(&self) {
|
||||||
let run = self.private_clone();
|
let run = self.private_clone();
|
||||||
|
|
||||||
|
if run.cluster.is_some() {
|
||||||
|
log::warn!("Run in a cluster");
|
||||||
|
} else {
|
||||||
|
log::warn!("Run not in a cluster");
|
||||||
|
}
|
||||||
|
|
||||||
// if we're a follower we starts a thread to register the tasks coming from the leader
|
// if we're a follower we starts a thread to register the tasks coming from the leader
|
||||||
if let Some(Cluster::Follower(follower)) = self.cluster.clone() {
|
if let Some(Cluster::Follower(follower)) = self.cluster.clone() {
|
||||||
let this = self.private_clone();
|
let this = self.private_clone();
|
||||||
|
if this.cluster.is_some() {
|
||||||
|
log::warn!("this in a cluster");
|
||||||
|
} else {
|
||||||
|
log::warn!("this not in a cluster");
|
||||||
|
}
|
||||||
std::thread::spawn(move || loop {
|
std::thread::spawn(move || loop {
|
||||||
let (task, content) = follower.write().unwrap().get_new_task();
|
let (task, content) = follower.get_new_task();
|
||||||
this.register_raw_task(task, content);
|
this.register_raw_task(task, content);
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
@ -917,7 +928,7 @@ impl IndexScheduler {
|
|||||||
} else {
|
} else {
|
||||||
None
|
None
|
||||||
};
|
};
|
||||||
leader.write().unwrap().register_new_task(task.clone(), update_file);
|
leader.register_new_task(task.clone(), update_file);
|
||||||
}
|
}
|
||||||
|
|
||||||
// If the registered task is a task cancelation
|
// If the registered task is a task cancelation
|
||||||
@ -1283,21 +1294,30 @@ impl IndexScheduler {
|
|||||||
/// If there is no cluster or if leader -> create a new batch
|
/// If there is no cluster or if leader -> create a new batch
|
||||||
/// If follower -> wait till the leader gives us a batch to process
|
/// If follower -> wait till the leader gives us a batch to process
|
||||||
fn get_or_create_next_batch(&self) -> Result<Option<Batch>> {
|
fn get_or_create_next_batch(&self) -> Result<Option<Batch>> {
|
||||||
let rtxn = self.env.read_txn().map_err(Error::HeedTransaction)?;
|
info!("inside get or create next batch");
|
||||||
|
|
||||||
let batch = match &self.cluster {
|
let batch = match &self.cluster {
|
||||||
None | Some(Cluster::Leader(_)) => {
|
None | Some(Cluster::Leader(_)) => {
|
||||||
|
let rtxn = self.env.read_txn().map_err(Error::HeedTransaction)?;
|
||||||
self.create_next_batch(&rtxn).map_err(|e| Error::CreateBatch(Box::new(e)))?
|
self.create_next_batch(&rtxn).map_err(|e| Error::CreateBatch(Box::new(e)))?
|
||||||
}
|
}
|
||||||
Some(Cluster::Follower(follower)) => {
|
Some(Cluster::Follower(follower)) => {
|
||||||
let batch = follower.write().unwrap().get_new_batch();
|
let batch = follower.get_new_batch();
|
||||||
Some(self.get_batch_from_cluster_batch(&rtxn, batch)?)
|
Some(self.get_batch_from_cluster_batch(batch)?)
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
if self.cluster.is_some() {
|
||||||
|
println!("HERE: Part of a cluster");
|
||||||
|
} else if self.cluster.is_none() {
|
||||||
|
println!("HERE: Not part of a cluster");
|
||||||
|
}
|
||||||
|
info!("before checking if i’m a leader");
|
||||||
if let Some(Cluster::Leader(leader)) = &self.cluster {
|
if let Some(Cluster::Leader(leader)) = &self.cluster {
|
||||||
|
info!("I'm a leader");
|
||||||
if let Some(ref batch) = batch {
|
if let Some(ref batch) = batch {
|
||||||
leader.write().unwrap().starts_batch(batch.clone().into());
|
info!("I'm a leader and I got a batch to process");
|
||||||
|
leader.starts_batch(batch.clone().into());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Ok(batch)
|
Ok(batch)
|
||||||
|
@ -31,7 +31,7 @@ use error::PayloadError;
|
|||||||
use extractors::payload::PayloadConfig;
|
use extractors::payload::PayloadConfig;
|
||||||
use http::header::CONTENT_TYPE;
|
use http::header::CONTENT_TYPE;
|
||||||
use index_scheduler::{Cluster, IndexScheduler, IndexSchedulerOptions};
|
use index_scheduler::{Cluster, IndexScheduler, IndexSchedulerOptions};
|
||||||
use log::error;
|
use log::{error, info};
|
||||||
use meilisearch_auth::AuthController;
|
use meilisearch_auth::AuthController;
|
||||||
use meilisearch_types::milli::documents::{DocumentsBatchBuilder, DocumentsBatchReader};
|
use meilisearch_types::milli::documents::{DocumentsBatchBuilder, DocumentsBatchReader};
|
||||||
use meilisearch_types::milli::update::{IndexDocumentsConfig, IndexDocumentsMethod};
|
use meilisearch_types::milli::update::{IndexDocumentsConfig, IndexDocumentsMethod};
|
||||||
@ -226,11 +226,13 @@ fn open_or_create_database_unchecked(
|
|||||||
let cluster = if let Some(ref cluster) = opt.cluster_configuration.experimental_enable_ha {
|
let cluster = if let Some(ref cluster) = opt.cluster_configuration.experimental_enable_ha {
|
||||||
match cluster.as_str() {
|
match cluster.as_str() {
|
||||||
"leader" => {
|
"leader" => {
|
||||||
|
info!("Starting as a leader");
|
||||||
let mut addr = opt.http_addr.to_socket_addrs().unwrap().next().unwrap();
|
let mut addr = opt.http_addr.to_socket_addrs().unwrap().next().unwrap();
|
||||||
addr.set_port(6666);
|
addr.set_port(6666);
|
||||||
Some(Cluster::Leader(Arc::new(RwLock::new(Leader::new(addr)))))
|
Some(Cluster::Leader(Leader::new(addr)))
|
||||||
}
|
}
|
||||||
"follower" => {
|
"follower" => {
|
||||||
|
info!("Starting as a follower");
|
||||||
let mut addr = opt
|
let mut addr = opt
|
||||||
.cluster_configuration
|
.cluster_configuration
|
||||||
.leader
|
.leader
|
||||||
@ -241,7 +243,7 @@ fn open_or_create_database_unchecked(
|
|||||||
.next()
|
.next()
|
||||||
.unwrap();
|
.unwrap();
|
||||||
addr.set_port(6666);
|
addr.set_port(6666);
|
||||||
Some(Cluster::Follower(Arc::new(RwLock::new(Follower::join(addr)))))
|
Some(Cluster::Follower(Follower::join(addr)))
|
||||||
}
|
}
|
||||||
_ => panic!("Available values for the cluster mode are leader and follower"),
|
_ => panic!("Available values for the cluster mode are leader and follower"),
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user