Skip to content

Commit df4db59

Browse files
committed
WIP: Implement LruCachingFileReader::new
1 parent 56beab3 commit df4db59

File tree

2 files changed

+10
-1
lines changed

2 files changed

+10
-1
lines changed

rust/cubestore/cubestore/src/queryplanner/metadata_cache.rs

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -136,6 +136,16 @@ pub struct LruCachingFileReader {
136136
cache: Arc<moka::sync::Cache<object_store::path::Path, Arc<ParquetMetaData>>>,
137137
}
138138

139+
impl LruCachingFileReader {
140+
pub fn new(path: object_store::path::Path, reader: Box<dyn AsyncFileReader>, cache: Arc<moka::sync::Cache<object_store::path::Path, Arc<ParquetMetaData>>>) -> LruCachingFileReader {
141+
LruCachingFileReader {
142+
path,
143+
reader,
144+
cache,
145+
}
146+
}
147+
}
148+
139149
impl AsyncFileReader for LruCachingFileReader {
140150
fn get_bytes(
141151
&mut self,

rust/cubestore/cubestore/src/queryplanner/partition_filter.rs

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,6 @@ impl PartitionFilter {
1919
const SIZE_LIMIT: usize = 50;
2020

2121
pub fn extract(s: &Schema, filters: &[Expr]) -> PartitionFilter {
22-
println!("Calling extract on filters {:?}", filters);
2322
let builder = Builder { schema: s };
2423

2524
let mut r = vec![];

0 commit comments

Comments
 (0)