diff --git a/parquet_file/src/serialize.rs b/parquet_file/src/serialize.rs index 052cde23b3..685d389074 100644 --- a/parquet_file/src/serialize.rs +++ b/parquet_file/src/serialize.rs @@ -15,8 +15,8 @@ use thiserror::Error; use crate::metadata::{IoxMetadata, METADATA_KEY}; -/// Parquet row group size -pub const ROW_GROUP_SIZE: usize = 1024 * 1024; +/// Parquet row group write size +pub const ROW_GROUP_WRITE_SIZE: usize = 1024 * 1024; /// [`RecordBatch`] to Parquet serialisation errors. #[derive(Debug, Error)] @@ -155,7 +155,7 @@ fn writer_props(meta: &IoxMetadata) -> Result. +#[allow(clippy::assertions_on_constants)] +const _: () = assert!(ROW_GROUP_WRITE_SIZE % ROW_GROUP_READ_SIZE == 0); /// Errors returned during a Parquet "put" operation, covering [`RecordBatch`] /// pull from the provided stream, encoding, and finally uploading the bytes to /// the object store. @@ -275,7 +282,8 @@ async fn download_and_scan_parquet( }; let mask = ProjectionMask::roots(arrow_reader.parquet_schema(), mask); - let record_batch_reader = arrow_reader.get_record_reader_by_columns(mask, ROW_GROUP_SIZE)?; + let record_batch_reader = + arrow_reader.get_record_reader_by_columns(mask, ROW_GROUP_READ_SIZE)?; for batch in record_batch_reader { let batch = batch.map(|batch| {