Skip to content

Commit c0494fe

Browse files
authored
feat: new crate trace_metric for collecting metrics in read procedure (#714)
* feat: add ReadMetricsCollector for collecting metrics in read procedure * collect metrics during merge * add new component trace_metric * add derive crate * impl proc macro for trace_metric * use the trace_metric crate * support optional collector * fix license * make collector is optional * trace metrics in parquet reader * rename Collector to MetricsCollector * rename TracedMetrics to TraceMetricWhenDrop * support FormatCollectorVisitor * set the metrics into datafusion * remove useless file * remove useless module * rename constant variable * fix wrong field name in metrics * address CR * attach index to the merge iter's metrics collector * rename some constant variables * fix wrong variable name * rename metric type
1 parent b1742f9 commit c0494fe

39 files changed

+762
-75
lines changed

Cargo.lock

+27
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

Cargo.toml

+6
Original file line numberDiff line numberDiff line change
@@ -29,6 +29,9 @@ members = [
2929
"components/profile",
3030
"components/skiplist",
3131
"components/table_kv",
32+
"components/trace_metric",
33+
"components/trace_metric_derive",
34+
"components/trace_metric_derive_tests",
3235
"components/tracing_util",
3336
"df_operator",
3437
"integration_tests",
@@ -112,6 +115,9 @@ table_engine = { path = "table_engine" }
112115
table_kv = { path = "components/table_kv" }
113116
tempfile = "3.1.0"
114117
tracing_util = { path = "components/tracing_util" }
118+
trace_metric = { path = "components/trace_metric" }
119+
trace_metric_derive = { path = "components/trace_metric_derive" }
120+
trace_metric_derive_tests = { path = "components/trace_metric_derive_tests" }
115121
tonic = "0.8.1"
116122
tokio = { version = "1.25", features = ["full"] }
117123
wal = { path = "wal" }

analytic_engine/Cargo.toml

+1
Original file line numberDiff line numberDiff line change
@@ -46,6 +46,7 @@ table_engine = { workspace = true }
4646
table_kv = { workspace = true }
4747
tempfile = { workspace = true, optional = true }
4848
tokio = { workspace = true }
49+
trace_metric = { workspace = true }
4950
wal = { workspace = true }
5051
xorfilter-rs = { workspace = true }
5152

analytic_engine/src/instance/flush_compaction.rs

+2
Original file line numberDiff line numberDiff line change
@@ -855,6 +855,7 @@ impl SpaceStore {
855855
let sequence = table_data.last_sequence();
856856
let mut builder = MergeBuilder::new(MergeConfig {
857857
request_id,
858+
metrics_collector: None,
858859
// no need to set deadline for compaction
859860
deadline: None,
860861
space_id,
@@ -1053,6 +1054,7 @@ fn build_mem_table_iter(memtable: MemTableRef, table_data: &TableData) -> Result
10531054
projected_schema: ProjectedSchema::no_projection(table_data.schema()),
10541055
need_dedup: table_data.dedup(),
10551056
reverse: false,
1057+
metrics_collector: None,
10561058
};
10571059
memtable
10581060
.scan(scan_ctx, scan_req)

analytic_engine/src/instance/read.rs

+26-8
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,7 @@ use table_engine::{
2323
table::ReadRequest,
2424
};
2525
use tokio::sync::mpsc::{self, Receiver};
26+
use trace_metric::Metric;
2627

2728
use crate::{
2829
instance::Instance,
@@ -66,6 +67,9 @@ pub enum Error {
6667
define_result!(Error);
6768

6869
const RECORD_BATCH_READ_BUF_SIZE: usize = 1000;
70+
const MERGE_SORT_METRIC_NAME: &str = "do_merge_sort";
71+
const ITER_NUM_METRIC_NAME: &str = "iter_num";
72+
const MERGE_ITER_METRICS_COLLECTOR_NAME_PREFIX: &str = "merge_iter";
6973

7074
/// Check whether it needs to apply merge sorting when reading the table with
7175
/// the `table_options` by the `read_request`.
@@ -91,13 +95,18 @@ impl Instance {
9195

9296
let table_data = space_table.table_data();
9397

94-
// Collect metrics.
95-
table_data.metrics.on_read_request_begin();
96-
9798
let iter_options = self.iter_options.clone();
9899
let table_options = table_data.table_options();
99100

100-
if need_merge_sort_streams(&table_data.table_options(), &request) {
101+
// Collect metrics.
102+
table_data.metrics.on_read_request_begin();
103+
let need_merge_sort = need_merge_sort_streams(&table_options, &request);
104+
request.metrics_collector.collect(Metric::boolean(
105+
MERGE_SORT_METRIC_NAME.to_string(),
106+
need_merge_sort,
107+
));
108+
109+
if need_merge_sort {
101110
let merge_iters = self
102111
.build_merge_iters(table_data, &request, iter_options, &table_options)
103112
.await?;
@@ -123,16 +132,16 @@ impl Instance {
123132
};
124133

125134
// Split iterators into `read_parallelism` groups.
126-
let mut splited_iters: Vec<_> = std::iter::repeat_with(Vec::new)
135+
let mut splitted_iters: Vec<_> = std::iter::repeat_with(Vec::new)
127136
.take(read_parallelism)
128137
.collect();
129138

130139
for (i, time_aligned_iter) in partitioned_iters.into_iter().enumerate() {
131-
splited_iters[i % read_parallelism].push(time_aligned_iter);
140+
splitted_iters[i % read_parallelism].push(time_aligned_iter);
132141
}
133142

134143
let mut streams = Vec::with_capacity(read_parallelism);
135-
for iters in splited_iters {
144+
for iters in splitted_iters {
136145
let stream = iters_to_stream(iters, self.read_runtime(), &request.projected_schema);
137146
streams.push(stream);
138147
}
@@ -169,9 +178,13 @@ impl Instance {
169178
let read_views = self.partition_ssts_and_memtables(time_range, version, table_options);
170179

171180
let mut iters = Vec::with_capacity(read_views.len());
172-
for read_view in read_views {
181+
for (idx, read_view) in read_views.into_iter().enumerate() {
182+
let metrics_collector = request
183+
.metrics_collector
184+
.span(format!("{MERGE_ITER_METRICS_COLLECTOR_NAME_PREFIX}_{idx}"));
173185
let merge_config = MergeConfig {
174186
request_id: request.request_id,
187+
metrics_collector: Some(metrics_collector),
175188
deadline: request.opts.deadline,
176189
space_id: table_data.space_id,
177190
table_id: table_data.id,
@@ -201,6 +214,11 @@ impl Instance {
201214
iters.push(dedup_iter);
202215
}
203216

217+
request.metrics_collector.collect(Metric::number(
218+
ITER_NUM_METRIC_NAME.to_string(),
219+
iters.len(),
220+
));
221+
204222
Ok(iters)
205223
}
206224

analytic_engine/src/memtable/mod.rs

+5-2
Original file line numberDiff line numberDiff line change
@@ -18,6 +18,7 @@ use common_types::{
1818
};
1919
use common_util::{define_result, error::GenericError};
2020
use snafu::{Backtrace, Snafu};
21+
use trace_metric::MetricsCollector;
2122

2223
use crate::memtable::key::KeySequence;
2324

@@ -129,14 +130,16 @@ pub struct ScanRequest {
129130
pub projected_schema: ProjectedSchema,
130131
pub need_dedup: bool,
131132
pub reverse: bool,
133+
/// Collector for scan metrics.
134+
pub metrics_collector: Option<MetricsCollector>,
132135
}
133136

134137
/// In memory storage for table's data.
135138
///
136139
/// # Concurrency
137-
/// The memtable is designed for single-writer and mutltiple-reader usage, so
140+
/// The memtable is designed for single-writer and multiple-reader usage, so
138141
/// not all function supports concurrent writer, the caller should guarantee not
139-
/// writing to the memtable concurrrently.
142+
/// writing to the memtable concurrently.
140143
// All operation is done in memory, no need to use async trait
141144
pub trait MemTable {
142145
/// Schema of this memtable

analytic_engine/src/memtable/skiplist/mod.rs

+4
Original file line numberDiff line numberDiff line change
@@ -204,6 +204,7 @@ mod tests {
204204
projected_schema: projected_schema.clone(),
205205
need_dedup: true,
206206
reverse: false,
207+
metrics_collector: None,
207208
},
208209
vec![
209210
build_row(b"a", 1, 10.0, "v1", 1000, 1_000_000),
@@ -223,6 +224,7 @@ mod tests {
223224
projected_schema: projected_schema.clone(),
224225
need_dedup: true,
225226
reverse: false,
227+
metrics_collector: None,
226228
},
227229
vec![
228230
build_row(b"a", 1, 10.0, "v1", 1000, 1_000_000),
@@ -241,6 +243,7 @@ mod tests {
241243
projected_schema,
242244
need_dedup: true,
243245
reverse: false,
246+
metrics_collector: None,
244247
},
245248
vec![
246249
build_row(b"a", 1, 10.0, "v1", 1000, 1_000_000),
@@ -272,6 +275,7 @@ mod tests {
272275
projected_schema,
273276
need_dedup: true,
274277
reverse: false,
278+
metrics_collector: None,
275279
},
276280
vec![
277281
build_row_for_two_column(b"a", 1),

analytic_engine/src/row_iter/chain.rs

+3
Original file line numberDiff line numberDiff line change
@@ -116,6 +116,7 @@ impl<'a> Builder<'a> {
116116
false,
117117
self.config.predicate.as_ref(),
118118
self.config.deadline,
119+
None,
119120
)
120121
.context(BuildStreamFromMemtable)?;
121122
streams.push(stream);
@@ -131,6 +132,7 @@ impl<'a> Builder<'a> {
131132
false,
132133
self.config.predicate.as_ref(),
133134
self.config.deadline,
135+
None,
134136
)
135137
.context(BuildStreamFromMemtable)?;
136138
streams.push(stream);
@@ -145,6 +147,7 @@ impl<'a> Builder<'a> {
145147
self.config.sst_factory,
146148
&self.config.sst_read_options,
147149
self.config.store_picker,
150+
None,
148151
)
149152
.await
150153
.context(BuildStreamFromSst)?;

0 commit comments

Comments
 (0)