mirror of
https://github.com/meilisearch/meilisearch.git
synced 2024-11-22 10:07:40 +08:00
makes the analytics works with the option and the feature
This commit is contained in:
parent
e226b1a87f
commit
664d09e86a
2
Cargo.lock
generated
2
Cargo.lock
generated
@ -2543,7 +2543,7 @@ dependencies = [
|
||||
[[package]]
|
||||
name = "segment"
|
||||
version = "0.1.1"
|
||||
source = "git+https://github.com/meilisearch/segment#656b91e1f7a2c6443e2a8ed59f8942400e9a811e"
|
||||
source = "git+https://github.com/meilisearch/segment#042a8631361f02ba84e8bb06f9120e93bf1922f2"
|
||||
dependencies = [
|
||||
"async-trait",
|
||||
"chrono",
|
||||
|
@ -1,164 +1,210 @@
|
||||
use meilisearch_lib::index_controller::Stats;
|
||||
use meilisearch_lib::MeiliSearch;
|
||||
use once_cell::sync::Lazy;
|
||||
use segment::message::{Identify, Track, User};
|
||||
use segment::{AutoBatcher, Batcher, HttpClient};
|
||||
use serde_json::{json, Value};
|
||||
use serde_json::Value;
|
||||
use std::fmt::Display;
|
||||
use std::time::{Duration, Instant};
|
||||
use sysinfo::DiskExt;
|
||||
use sysinfo::ProcessorExt;
|
||||
use sysinfo::System;
|
||||
use sysinfo::SystemExt;
|
||||
use tokio::sync::Mutex;
|
||||
use uuid::Uuid;
|
||||
|
||||
use crate::Opt;
|
||||
|
||||
const SEGMENT_API_KEY: &str = "vHi89WrNDckHSQssyUJqLvIyp2QFITSC";
|
||||
// if we are in release mode and the feature analytics was enabled
|
||||
#[cfg(all(not(debug_assertions), feature = "analytics"))]
|
||||
mod segment {
|
||||
use crate::analytics::Analytics;
|
||||
use meilisearch_lib::index_controller::Stats;
|
||||
use meilisearch_lib::MeiliSearch;
|
||||
use once_cell::sync::Lazy;
|
||||
use segment::message::{Identify, Track, User};
|
||||
use segment::{AutoBatcher, Batcher, HttpClient};
|
||||
use serde_json::{json, Value};
|
||||
use std::fmt::Display;
|
||||
use std::time::{Duration, Instant};
|
||||
use sysinfo::DiskExt;
|
||||
use sysinfo::ProcessorExt;
|
||||
use sysinfo::System;
|
||||
use sysinfo::SystemExt;
|
||||
use tokio::sync::Mutex;
|
||||
use uuid::Uuid;
|
||||
|
||||
pub struct Analytics {
|
||||
user: User,
|
||||
opt: Opt,
|
||||
batcher: Mutex<AutoBatcher>,
|
||||
}
|
||||
use crate::Opt;
|
||||
|
||||
impl Analytics {
|
||||
pub fn publish(&'static self, event_name: String, send: Value) {
|
||||
tokio::spawn(async move {
|
||||
let _ = self
|
||||
const SEGMENT_API_KEY: &str = "vHi89WrNDckHSQssyUJqLvIyp2QFITSC";
|
||||
|
||||
pub struct SegmentAnalytics {
|
||||
user: User,
|
||||
opt: Opt,
|
||||
batcher: Mutex<AutoBatcher>,
|
||||
}
|
||||
|
||||
impl SegmentAnalytics {
|
||||
fn compute_traits(opt: &Opt, stats: Stats) -> Value {
|
||||
static FIRST_START_TIMESTAMP: Lazy<Instant> = Lazy::new(Instant::now);
|
||||
static SYSTEM: Lazy<Value> = Lazy::new(|| {
|
||||
let mut sys = System::new_all();
|
||||
sys.refresh_all();
|
||||
json!({
|
||||
"distribution": sys.name().zip(sys.kernel_version()).map(|(name, version)| format!("{}: {}", name, version)),
|
||||
"core_number": sys.processors().len(),
|
||||
"ram_size": sys.total_memory(),
|
||||
"frequency": sys.processors().iter().map(|cpu| cpu.frequency()).sum::<u64>() / sys.processors().len() as u64,
|
||||
"disk_size": sys.disks().iter().map(|disk| disk.available_space()).max(),
|
||||
"server_provider": std::env::var("MEILI_SERVER_PROVIDER").ok(),
|
||||
})
|
||||
});
|
||||
let number_of_documents = stats
|
||||
.indexes
|
||||
.values()
|
||||
.map(|index| index.number_of_documents)
|
||||
.collect::<Vec<u64>>();
|
||||
|
||||
json!({
|
||||
"system": *SYSTEM,
|
||||
"stats": {
|
||||
"database_size": stats.database_size,
|
||||
"indexes_number": stats.indexes.len(),
|
||||
"documents_number": number_of_documents,
|
||||
},
|
||||
"infos": {
|
||||
"version": env!("CARGO_PKG_VERSION").to_string(),
|
||||
"env": opt.env.clone(),
|
||||
"snapshot": opt.schedule_snapshot,
|
||||
"start_since_days": FIRST_START_TIMESTAMP.elapsed().as_secs() / 60 * 60 * 24, // one day
|
||||
},
|
||||
})
|
||||
}
|
||||
|
||||
pub async fn new(opt: &Opt, meilisearch: &MeiliSearch) -> &'static Self {
|
||||
// see if there is already a user-id
|
||||
let user_id = std::fs::read_to_string(opt.db_path.join("user-id"));
|
||||
let first_time_run = user_id.is_err();
|
||||
// if not, generate a new user-id and save it to the fs
|
||||
let user_id = user_id.unwrap_or_else(|_| Uuid::new_v4().to_string());
|
||||
let _ = std::fs::write(opt.db_path.join("user-id"), user_id.as_bytes());
|
||||
|
||||
let client = HttpClient::default();
|
||||
let user = User::UserId {
|
||||
user_id: user_id.clone(),
|
||||
};
|
||||
let batcher = Mutex::new(AutoBatcher::new(
|
||||
client,
|
||||
Batcher::new(None),
|
||||
SEGMENT_API_KEY.to_string(),
|
||||
));
|
||||
let segment = Box::new(Self {
|
||||
user,
|
||||
opt: opt.clone(),
|
||||
batcher,
|
||||
});
|
||||
let segment = Box::leak(segment);
|
||||
|
||||
// send an identify event
|
||||
let _ = segment
|
||||
.batcher
|
||||
.lock()
|
||||
.await
|
||||
.push(Track {
|
||||
user: self.user.clone(),
|
||||
event: event_name.clone(),
|
||||
properties: send,
|
||||
.push(Identify {
|
||||
user: segment.user.clone(),
|
||||
// TODO: TAMO: what should we do when meilisearch is broken at start
|
||||
traits: Self::compute_traits(
|
||||
&segment.opt,
|
||||
meilisearch.get_all_stats().await.unwrap(),
|
||||
),
|
||||
..Default::default()
|
||||
})
|
||||
.await;
|
||||
println!("ANALYTICS: {} added to batch", event_name)
|
||||
});
|
||||
}
|
||||
|
||||
pub fn tick(&'static self, meilisearch: MeiliSearch) {
|
||||
tokio::spawn(async move {
|
||||
loop {
|
||||
tokio::time::sleep(Duration::from_secs(60)).await; // 1 minutes
|
||||
println!("ANALYTICS: should do things");
|
||||
|
||||
if let Ok(stats) = meilisearch.get_all_stats().await {
|
||||
let traits = Self::compute_traits(&self.opt, stats);
|
||||
let user = self.user.clone();
|
||||
println!("ANALYTICS: Pushing our identify tick");
|
||||
let _ = self
|
||||
.batcher
|
||||
.lock()
|
||||
.await
|
||||
.push(Identify {
|
||||
user,
|
||||
traits,
|
||||
..Default::default()
|
||||
})
|
||||
.await;
|
||||
}
|
||||
println!("ANALYTICS: Pushing our batch");
|
||||
let _ = self.batcher.lock().await.flush().await;
|
||||
// send the associated track event
|
||||
if first_time_run {
|
||||
segment.publish("Launched for the first time".to_string(), json!({}));
|
||||
}
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
impl Analytics {
|
||||
pub async fn new(opt: &Opt, meilisearch: &MeiliSearch) -> &'static Self {
|
||||
let user_id = std::fs::read_to_string(opt.db_path.join("user-id"));
|
||||
let first_time_run = user_id.is_err();
|
||||
let user_id = user_id.unwrap_or(Uuid::new_v4().to_string());
|
||||
let _ = std::fs::write(opt.db_path.join("user-id"), user_id.as_bytes());
|
||||
let client = HttpClient::default();
|
||||
let user = User::UserId {
|
||||
user_id: user_id.clone(),
|
||||
};
|
||||
let batcher = Batcher::new(None);
|
||||
let batcher = Mutex::new(AutoBatcher::new(
|
||||
client,
|
||||
batcher,
|
||||
SEGMENT_API_KEY.to_string(),
|
||||
));
|
||||
let segment = Box::new(Self {
|
||||
user,
|
||||
opt: opt.clone(),
|
||||
batcher,
|
||||
});
|
||||
let segment = Box::leak(segment);
|
||||
// start the runtime tick
|
||||
segment.tick(meilisearch.clone());
|
||||
|
||||
// send an identify event
|
||||
let _ = segment
|
||||
.batcher
|
||||
.lock()
|
||||
.await
|
||||
.push(Identify {
|
||||
user: segment.user.clone(),
|
||||
// TODO: TAMO: what should we do when meilisearch is broken at start
|
||||
traits: Self::compute_traits(
|
||||
&segment.opt,
|
||||
meilisearch.get_all_stats().await.unwrap(),
|
||||
),
|
||||
..Default::default()
|
||||
})
|
||||
.await;
|
||||
println!("ANALYTICS: pushed the identify event");
|
||||
|
||||
// send the associated track event
|
||||
if first_time_run {
|
||||
segment.publish("Launched for the first time".to_string(), json!({}));
|
||||
segment
|
||||
}
|
||||
|
||||
// start the runtime tick
|
||||
segment.tick(meilisearch.clone());
|
||||
fn tick(&'static self, meilisearch: MeiliSearch) {
|
||||
tokio::spawn(async move {
|
||||
loop {
|
||||
tokio::time::sleep(Duration::from_secs(60)).await; // 1 minutes
|
||||
println!("ANALYTICS: should do things");
|
||||
|
||||
segment
|
||||
if let Ok(stats) = meilisearch.get_all_stats().await {
|
||||
let traits = Self::compute_traits(&self.opt, stats);
|
||||
let user = self.user.clone();
|
||||
println!("ANALYTICS: Pushing our identify tick");
|
||||
let _ = self
|
||||
.batcher
|
||||
.lock()
|
||||
.await
|
||||
.push(Identify {
|
||||
user,
|
||||
traits,
|
||||
..Default::default()
|
||||
})
|
||||
.await;
|
||||
}
|
||||
let _ = self.batcher.lock().await.flush().await;
|
||||
}
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
fn compute_traits(opt: &Opt, stats: Stats) -> Value {
|
||||
static FIRST_START_TIMESTAMP: Lazy<Instant> = Lazy::new(|| Instant::now());
|
||||
static SYSTEM: Lazy<Value> = Lazy::new(|| {
|
||||
let mut sys = System::new_all();
|
||||
sys.refresh_all();
|
||||
json!({
|
||||
"distribution": sys.name().zip(sys.kernel_version()).map(|(name, version)| format!("{}: {}", name, version)),
|
||||
"core_number": sys.processors().len(),
|
||||
"ram_size": sys.total_memory(),
|
||||
"frequency": sys.processors().iter().map(|cpu| cpu.frequency()).sum::<u64>() / sys.processors().len() as u64,
|
||||
"disk_size": sys.disks().iter().map(|disk| disk.available_space()).max(),
|
||||
"server_provider": std::env::var("MEILI_SERVER_PROVIDER").ok(),
|
||||
})
|
||||
});
|
||||
let number_of_documents = stats
|
||||
.indexes
|
||||
.values()
|
||||
.map(|index| index.number_of_documents)
|
||||
.collect::<Vec<u64>>();
|
||||
#[async_trait::async_trait]
|
||||
impl super::Analytics for SegmentAnalytics {
|
||||
fn publish(&'static self, event_name: String, send: Value) {
|
||||
tokio::spawn(async move {
|
||||
let _ = self
|
||||
.batcher
|
||||
.lock()
|
||||
.await
|
||||
.push(Track {
|
||||
user: self.user.clone(),
|
||||
event: event_name.clone(),
|
||||
properties: send,
|
||||
..Default::default()
|
||||
})
|
||||
.await;
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
json!({
|
||||
"system": *SYSTEM,
|
||||
"stats": {
|
||||
"database_size": stats.database_size,
|
||||
"indexes_number": stats.indexes.len(),
|
||||
"documents_number": number_of_documents,
|
||||
},
|
||||
"infos": {
|
||||
"version": env!("CARGO_PKG_VERSION").to_string(),
|
||||
"env": opt.env.clone(),
|
||||
"snapshot": opt.schedule_snapshot,
|
||||
"start_since_days": FIRST_START_TIMESTAMP.elapsed().as_secs() / 60 * 60 * 24, // one day
|
||||
},
|
||||
})
|
||||
impl Display for SegmentAnalytics {
|
||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||
write!(f, "{}", self.user)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl Display for Analytics {
|
||||
// if we are in debug mode OR the analytics feature is disabled
|
||||
#[cfg(any(debug_assertions, not(feature = "analytics")))]
|
||||
pub type SegmentAnalytics = MockAnalytics;
|
||||
#[cfg(all(not(debug_assertions), feature = "analytics"))]
|
||||
pub type SegmentAnalytics = segment::SegmentAnalytics;
|
||||
|
||||
pub struct MockAnalytics {
|
||||
user: String,
|
||||
}
|
||||
|
||||
impl MockAnalytics {
|
||||
pub fn new(opt: &Opt) -> &'static Self {
|
||||
let user = std::fs::read_to_string(opt.db_path.join("user-id"))
|
||||
.unwrap_or_else(|_| "No user-id".to_string());
|
||||
let analytics = Box::new(Self { user });
|
||||
Box::leak(analytics)
|
||||
}
|
||||
}
|
||||
|
||||
#[async_trait::async_trait]
|
||||
impl Analytics for MockAnalytics {
|
||||
/// This is a noop and should be optimized out
|
||||
fn publish(&'static self, _event_name: String, _send: Value) {}
|
||||
}
|
||||
|
||||
impl Display for MockAnalytics {
|
||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||
write!(f, "{:?}", self.user)
|
||||
write!(f, "{}", self.user)
|
||||
}
|
||||
}
|
||||
|
||||
#[async_trait::async_trait]
|
||||
pub trait Analytics: Display {
|
||||
fn publish(&'static self, event_name: String, send: Value);
|
||||
}
|
||||
|
@ -3,7 +3,6 @@
|
||||
pub mod error;
|
||||
#[macro_use]
|
||||
pub mod extractors;
|
||||
#[cfg(all(not(debug_assertions), feature = "analytics"))]
|
||||
pub mod analytics;
|
||||
pub mod helpers;
|
||||
pub mod option;
|
||||
|
@ -1,13 +1,12 @@
|
||||
use std::env;
|
||||
|
||||
use actix_web::HttpServer;
|
||||
use meilisearch_http::analytics;
|
||||
use meilisearch_http::analytics::Analytics;
|
||||
use meilisearch_http::{create_app, setup_meilisearch, Opt};
|
||||
use meilisearch_lib::MeiliSearch;
|
||||
use structopt::StructOpt;
|
||||
|
||||
#[cfg(all(not(debug_assertions), feature = "analytics"))]
|
||||
use meilisearch_http::analytics;
|
||||
|
||||
#[cfg(target_os = "linux")]
|
||||
#[global_allocator]
|
||||
static ALLOC: tikv_jemallocator::Jemalloc = tikv_jemallocator::Jemalloc;
|
||||
@ -47,12 +46,15 @@ async fn main() -> anyhow::Result<()> {
|
||||
let meilisearch = setup_meilisearch(&opt)?;
|
||||
|
||||
#[cfg(all(not(debug_assertions), feature = "analytics"))]
|
||||
if !opt.no_analytics {
|
||||
let analytics = analytics::Analytics::new(&opt, &meilisearch).await;
|
||||
println!("go my analytics back");
|
||||
}
|
||||
let analytics = if !opt.no_analytics {
|
||||
analytics::SegmentAnalytics::new(&opt, &meilisearch).await as &'static dyn Analytics
|
||||
} else {
|
||||
analytics::MockAnalytics::new(&opt) as &'static dyn Analytics
|
||||
};
|
||||
#[cfg(any(debug_assertions, not(feature = "analytics")))]
|
||||
let analytics = analytics::MockAnalytics::new(&opt);
|
||||
|
||||
print_launch_resume(&opt);
|
||||
print_launch_resume(&opt, analytics);
|
||||
|
||||
run_http(meilisearch, opt).await?;
|
||||
|
||||
@ -77,7 +79,7 @@ async fn run_http(data: MeiliSearch, opt: Opt) -> anyhow::Result<()> {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
pub fn print_launch_resume(opt: &Opt) {
|
||||
pub fn print_launch_resume(opt: &Opt, analytics: &'static dyn Analytics) {
|
||||
let commit_sha = option_env!("VERGEN_GIT_SHA").unwrap_or("unknown");
|
||||
let commit_date = option_env!("VERGEN_GIT_COMMIT_TIMESTAMP").unwrap_or("unknown");
|
||||
|
||||
@ -119,6 +121,7 @@ Anonymous telemetry: \"Enabled\""
|
||||
);
|
||||
}
|
||||
}
|
||||
eprintln!("Unique User ID:\t\"{}\"", analytics);
|
||||
|
||||
eprintln!();
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user