Skip to content
Merged
Show file tree
Hide file tree
Changes from 18 commits
Commits
Show all changes
46 commits
Select commit Hold shift + click to select a range
1f93a93
update
XiangpengHao Jul 1, 2025
2e01e56
update
XiangpengHao Jul 1, 2025
0bd08c3
update
XiangpengHao Jul 1, 2025
d6ecbd4
update
XiangpengHao Jul 1, 2025
7cd5518
cleanup
XiangpengHao Jul 2, 2025
4520048
update
XiangpengHao Jul 2, 2025
e6281bc
update
XiangpengHao Jul 2, 2025
6b6d4fc
update
XiangpengHao Jul 2, 2025
b696b66
update
XiangpengHao Jul 2, 2025
f60581f
update
XiangpengHao Jul 2, 2025
1851f0b
clippy and license
XiangpengHao Jul 2, 2025
5e414a8
Merge remote-tracking branch 'apache/main' into pushdown-v4
alamb Jul 7, 2025
58add51
bug fix
XiangpengHao Jul 8, 2025
470cc01
Merge remote-tracking branch 'refs/remotes/origin/pushdown-v3' into p…
XiangpengHao Jul 8, 2025
2bf3d38
clippy
XiangpengHao Jul 8, 2025
2cf1a8f
bug fix
XiangpengHao Jul 8, 2025
86e149c
switch to boolean array for row selection
XiangpengHao Jul 15, 2025
4d24172
Merge remote-tracking branch 'apache/main' into pushdown-v4
alamb Jul 15, 2025
be134d6
Add comments (OCD) and rename some fields
alamb Jul 15, 2025
eecaf99
Merge pull request #4 from alamb/alamb/pushdown_suggestions
XiangpengHao Jul 15, 2025
5537bcb
fmt
XiangpengHao Jul 15, 2025
b835163
fmt
alamb Jul 16, 2025
5132de8
Simplify projection caching
alamb Jul 16, 2025
253dad3
Move cache options construction to ArrayReaderBuilder, add builders
alamb Jul 16, 2025
5d9781e
update memory accounting
XiangpengHao Jul 17, 2025
2e20902
Merge remote-tracking branch 'refs/remotes/origin/pushdown-v4' into p…
XiangpengHao Jul 17, 2025
721d00c
Merge pull request #5 from alamb/alamb/simplify_cache
XiangpengHao Jul 17, 2025
f8aed80
Merge pull request #6 from alamb/alamb/cleaner_api
XiangpengHao Jul 17, 2025
884b591
update
XiangpengHao Jul 17, 2025
4f6b918
array size
XiangpengHao Jul 17, 2025
6c53bfd
add test case
XiangpengHao Jul 17, 2025
8ebe579
fix bug
XiangpengHao Jul 17, 2025
c240a52
clippy & fmt
XiangpengHao Jul 17, 2025
30a0d1c
Add config option for predicate cache memory limit
alamb Jul 23, 2025
ed3ce13
Add option to control predicate cache, documentation, ArrowReaderMetr…
alamb Jul 23, 2025
42d5520
Update parquet/src/arrow/arrow_reader/mod.rs
alamb Jul 24, 2025
6e618b3
Merge pull request #7 from alamb/alamb/test_memory_limit
XiangpengHao Jul 24, 2025
f70e46a
Clarify in documentation that cache is only for async decoder
alamb Jul 25, 2025
15d6826
add comment
alamb Jul 25, 2025
bec6d9c
Revert backwards incompatible changes to the Parquet reader API
alamb Jul 25, 2025
3e05cb2
Merge pull request #9 from alamb/alamb/revert_api_changes
XiangpengHao Jul 25, 2025
4d64dc0
Merge pull request #8 from alamb/alamb/pushdown-v4-cleanup
XiangpengHao Jul 25, 2025
8da582b
Merge remote-tracking branch 'apache/main' into pushdown-v4
alamb Aug 6, 2025
315e463
exclude nested column from cache
XiangpengHao Aug 7, 2025
1db701a
only use expanded selection when the column is one of cache column
XiangpengHao Aug 7, 2025
bea4433
Merge remote-tracking branch 'upstream/main' into pushdown-v4
XiangpengHao Aug 7, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
80 changes: 67 additions & 13 deletions parquet/src/arrow/array_reader/builder.rs
Original file line number Diff line number Diff line change
Expand Up @@ -15,13 +15,16 @@
// specific language governing permissions and limitations
// under the License.

use std::sync::Arc;
use std::sync::{Arc, Mutex};

use arrow_schema::{DataType, Fields, SchemaBuilder};

use crate::arrow::array_reader::byte_view_array::make_byte_view_array_reader;
use crate::arrow::array_reader::cached_array_reader::CacheRole;
use crate::arrow::array_reader::cached_array_reader::CachedArrayReader;
use crate::arrow::array_reader::empty_array::make_empty_array_reader;
use crate::arrow::array_reader::fixed_len_byte_array::make_fixed_len_byte_array_reader;
use crate::arrow::array_reader::row_group_cache::RowGroupCache;
use crate::arrow::array_reader::{
make_byte_array_dictionary_reader, make_byte_array_reader, ArrayReader,
FixedSizeListArrayReader, ListArrayReader, MapArrayReader, NullArrayReader,
Expand All @@ -39,6 +42,14 @@ pub struct ArrayReaderBuilder<'a> {
row_groups: &'a dyn RowGroups,
}

/// Cache options containing projection mask, cache, and role
#[derive(Clone)]
pub struct CacheOptions<'a> {
pub projection_mask: &'a ProjectionMask,
pub cache: Arc<Mutex<RowGroupCache>>,
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Practically there's no contention because there's not parallelism in decoding one row group. we add mutex here because we need to use Arc.

pub role: CacheRole,
}

impl<'a> ArrayReaderBuilder<'a> {
pub fn new(row_groups: &'a dyn RowGroups) -> Self {
Self { row_groups }
Expand All @@ -51,7 +62,25 @@ impl<'a> ArrayReaderBuilder<'a> {
mask: &ProjectionMask,
) -> Result<Box<dyn ArrayReader>> {
let reader = field
.and_then(|field| self.build_reader(field, mask).transpose())
.and_then(|field| self.build_reader(field, mask, None).transpose())
.transpose()?
.unwrap_or_else(|| make_empty_array_reader(self.num_rows()));

Ok(reader)
}

/// Create [`ArrayReader`] from parquet schema, projection mask, and parquet file reader.
pub fn build_array_reader_with_cache(
&self,
field: Option<&ParquetField>,
mask: &ProjectionMask,
cache_options: CacheOptions,
) -> Result<Box<dyn ArrayReader>> {
let reader = field
.and_then(|field| {
self.build_reader(field, mask, Some(cache_options))
.transpose()
})
.transpose()?
.unwrap_or_else(|| make_empty_array_reader(self.num_rows()));

Expand All @@ -67,15 +96,36 @@ impl<'a> ArrayReaderBuilder<'a> {
&self,
field: &ParquetField,
mask: &ProjectionMask,
cache_options: Option<CacheOptions>,
) -> Result<Option<Box<dyn ArrayReader>>> {
match field.field_type {
ParquetFieldType::Primitive { .. } => self.build_primitive_reader(field, mask),
ParquetFieldType::Primitive { col_idx, .. } => {
let Some(reader) = self.build_primitive_reader(field, mask)? else {
return Ok(None);
};
let Some(cache_options) = cache_options else {
return Ok(Some(reader));
};

if cache_options.projection_mask.leaf_included(col_idx) {
Ok(Some(Box::new(CachedArrayReader::new(
reader,
cache_options.cache,
col_idx,
cache_options.role,
))))
} else {
Ok(Some(reader))
}
}
ParquetFieldType::Group { .. } => match &field.arrow_type {
DataType::Map(_, _) => self.build_map_reader(field, mask),
DataType::Struct(_) => self.build_struct_reader(field, mask),
DataType::List(_) => self.build_list_reader(field, mask, false),
DataType::LargeList(_) => self.build_list_reader(field, mask, true),
DataType::FixedSizeList(_, _) => self.build_fixed_size_list_reader(field, mask),
DataType::Map(_, _) => self.build_map_reader(field, mask, cache_options),
DataType::Struct(_) => self.build_struct_reader(field, mask, cache_options),
DataType::List(_) => self.build_list_reader(field, mask, cache_options, false),
DataType::LargeList(_) => self.build_list_reader(field, mask, cache_options, true),
DataType::FixedSizeList(_, _) => {
self.build_fixed_size_list_reader(field, mask, cache_options)
}
d => unimplemented!("reading group type {} not implemented", d),
},
}
Expand All @@ -86,12 +136,13 @@ impl<'a> ArrayReaderBuilder<'a> {
&self,
field: &ParquetField,
mask: &ProjectionMask,
cache_options: Option<CacheOptions>,
) -> Result<Option<Box<dyn ArrayReader>>> {
let children = field.children().unwrap();
assert_eq!(children.len(), 2);

let key_reader = self.build_reader(&children[0], mask)?;
let value_reader = self.build_reader(&children[1], mask)?;
let key_reader = self.build_reader(&children[0], mask, cache_options.clone())?;
let value_reader = self.build_reader(&children[1], mask, cache_options)?;

match (key_reader, value_reader) {
(Some(key_reader), Some(value_reader)) => {
Expand Down Expand Up @@ -137,12 +188,13 @@ impl<'a> ArrayReaderBuilder<'a> {
&self,
field: &ParquetField,
mask: &ProjectionMask,
cache_options: Option<CacheOptions>,
is_large: bool,
) -> Result<Option<Box<dyn ArrayReader>>> {
let children = field.children().unwrap();
assert_eq!(children.len(), 1);

let reader = match self.build_reader(&children[0], mask)? {
let reader = match self.build_reader(&children[0], mask, cache_options)? {
Some(item_reader) => {
// Need to retrieve underlying data type to handle projection
let item_type = item_reader.get_data_type().clone();
Expand Down Expand Up @@ -184,11 +236,12 @@ impl<'a> ArrayReaderBuilder<'a> {
&self,
field: &ParquetField,
mask: &ProjectionMask,
cache_options: Option<CacheOptions>,
) -> Result<Option<Box<dyn ArrayReader>>> {
let children = field.children().unwrap();
assert_eq!(children.len(), 1);

let reader = match self.build_reader(&children[0], mask)? {
let reader = match self.build_reader(&children[0], mask, cache_options)? {
Some(item_reader) => {
let item_type = item_reader.get_data_type().clone();
let reader = match &field.arrow_type {
Expand Down Expand Up @@ -318,6 +371,7 @@ impl<'a> ArrayReaderBuilder<'a> {
&self,
field: &ParquetField,
mask: &ProjectionMask,
cache_options: Option<CacheOptions>,
) -> Result<Option<Box<dyn ArrayReader>>> {
let arrow_fields = match &field.arrow_type {
DataType::Struct(children) => children,
Expand All @@ -330,7 +384,7 @@ impl<'a> ArrayReaderBuilder<'a> {
let mut builder = SchemaBuilder::with_capacity(children.len());

for (arrow, parquet) in arrow_fields.iter().zip(children) {
if let Some(reader) = self.build_reader(parquet, mask)? {
if let Some(reader) = self.build_reader(parquet, mask, cache_options.clone())? {
// Need to retrieve underlying data type to handle projection
let child_type = reader.get_data_type().clone();
builder.push(arrow.as_ref().clone().with_data_type(child_type));
Expand Down
Loading
Loading