Skip to content

Commit 7196136

Browse files
committed
remove unwrap
1 parent 72d0db0 commit 7196136

File tree

2 files changed

+13
-24
lines changed

2 files changed

+13
-24
lines changed

analytic_engine/src/sst/parquet/async_reader.rs

+13-10
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,8 @@ use common_util::{
2323
time::InstantExt,
2424
};
2525
use datafusion::physical_plan::{
26-
file_format::ParquetFileMetrics, metrics::ExecutionPlanMetricsSet,
26+
file_format::{parquet::page_filter::PagePruningPredicate, ParquetFileMetrics},
27+
metrics::ExecutionPlanMetricsSet,
2728
};
2829
use futures::{future::BoxFuture, FutureExt, Stream, StreamExt, TryFutureExt};
2930
use log::{debug, error, info, warn};
@@ -211,22 +212,24 @@ impl<'a> Reader<'a> {
211212
);
212213

213214
let mut streams = Vec::with_capacity(target_row_group_chunks.len());
214-
let exprs =
215-
datafusion::optimizer::utils::conjunction(self.predicate.exprs().to_vec()).unwrap();
215+
let exprs = datafusion::optimizer::utils::conjunction(self.predicate.exprs().to_vec());
216216
let metrics_set = ExecutionPlanMetricsSet::new();
217-
let metrics = ParquetFileMetrics::new(1, "abc", &metrics_set);
217+
let metrics = ParquetFileMetrics::new(1, &self.path.to_string(), &metrics_set);
218218
for chunk in target_row_group_chunks {
219219
let object_store_reader =
220220
ObjectStoreReader::new(self.store.clone(), self.path.clone(), meta_data.clone());
221-
let page_predicate =
222-
datafusion::physical_plan::file_format::parquet::page_filter::PagePruningPredicate::try_new(&exprs, arrow_schema.clone()).context(DataFusionError)?;
223-
let row_selection = page_predicate
224-
.prune(&chunk, parquet_metadata, &metrics)
225-
.context(DataFusionError)?;
221+
let row_selection = if let Some(exprs) = &exprs {
222+
let page_predicate = PagePruningPredicate::try_new(&exprs, arrow_schema.clone())
223+
.context(DataFusionError)?;
224+
page_predicate
225+
.prune(&chunk, parquet_metadata, &metrics)
226+
.context(DataFusionError)?
227+
} else {
228+
None
229+
};
226230
let mut builder = ParquetRecordBatchStreamBuilder::new(object_store_reader)
227231
.await
228232
.with_context(|| ParquetError)?;
229-
230233
if let Some(selection) = row_selection {
231234
builder = builder.with_row_selection(selection);
232235
};

docs/minimal.toml

-14
Original file line numberDiff line numberDiff line change
@@ -16,17 +16,3 @@ data_dir = "/tmp/ceresdb"
1616
[analytic.wal]
1717
type = "RocksDB"
1818
data_dir = "/tmp/ceresdb"
19-
20-
[cluster_deployment]
21-
mode = "NoMeta"
22-
23-
[[cluster_deployment.topology.schema_shards]]
24-
schema = 'public'
25-
auto_create_tables = true
26-
27-
[[cluster_deployment.topology.schema_shards.shard_views]]
28-
shard_id = 0
29-
30-
[cluster_deployment.topology.schema_shards.shard_views.endpoint]
31-
addr = "127.0.0.1"
32-
port = 8831

0 commit comments

Comments
 (0)