diff --git a/ioxd_querier/src/lib.rs b/ioxd_querier/src/lib.rs index 99eb72439c..d9eb143f54 100644 --- a/ioxd_querier/src/lib.rs +++ b/ioxd_querier/src/lib.rs @@ -14,7 +14,6 @@ use ioxd_common::{ }; use metric::Registry; use object_store::DynObjectStore; -use parquet_file::storage::ParquetStorage; use querier::{ create_ingester_connections_by_shard, QuerierCatalogCache, QuerierDatabase, QuerierHandler, QuerierHandlerImpl, QuerierServer, @@ -182,7 +181,6 @@ pub async fn create_querier_server_type( QuerierDatabase::new( catalog_cache, Arc::clone(&args.metric_registry), - ParquetStorage::new(args.object_store), args.exec, ingester_connection, args.querier_config.max_concurrent_queries(), diff --git a/ioxd_querier/src/rpc/namespace.rs b/ioxd_querier/src/rpc/namespace.rs index f355342ccf..7c03c16401 100644 --- a/ioxd_querier/src/rpc/namespace.rs +++ b/ioxd_querier/src/rpc/namespace.rs @@ -56,7 +56,6 @@ mod tests { use super::*; use generated_types::influxdata::iox::namespace::v1::namespace_service_server::NamespaceService; use iox_tests::util::TestCatalog; - use parquet_file::storage::ParquetStorage; use querier::{create_ingester_connection_for_testing, QuerierCatalogCache}; use tokio::runtime::Handle; @@ -78,7 +77,6 @@ mod tests { QuerierDatabase::new( catalog_cache, catalog.metric_registry(), - ParquetStorage::new(catalog.object_store()), catalog.exec(), Some(create_ingester_connection_for_testing()), QuerierDatabase::MAX_CONCURRENT_QUERIES_MAX, @@ -115,7 +113,6 @@ mod tests { QuerierDatabase::new( catalog_cache, catalog.metric_registry(), - ParquetStorage::new(catalog.object_store()), catalog.exec(), Some(create_ingester_connection_for_testing()), QuerierDatabase::MAX_CONCURRENT_QUERIES_MAX, diff --git a/querier/src/chunk/mod.rs b/querier/src/chunk/mod.rs index 137aeb85a2..8f7197d669 100644 --- a/querier/src/chunk/mod.rs +++ b/querier/src/chunk/mod.rs @@ -188,6 +188,7 @@ impl From> for ChunkStage { #[derive(Debug, Clone, Copy, PartialEq, Eq, Default)] pub enum QuerierChunkLoadSetting { /// Only stay in parquet mode and never use read buffer data. + #[default] ParquetOnly, /// Only use read buffer data. @@ -195,12 +196,11 @@ pub enum QuerierChunkLoadSetting { /// This forces the querier to load the read buffer for this chunk. ReadBufferOnly, - /// Default "on-demand" handling. + /// "On-demand" handling. /// /// When the chunk is created, it will look up if there is already read buffer data loaded (or loading is already in /// progress) and is that. Otherwise it will parquet. If the actual row data is requested (via /// [`QueryChunk::read_filter`](iox_query::QueryChunk::read_filter)) then it will force-load the read buffer. - #[default] OnDemand, } @@ -358,10 +358,10 @@ impl ChunkAdapter { /// Create new adapter with empty cache. pub fn new( catalog_cache: Arc, - store: ParquetStorage, metric_registry: Arc, load_settings: HashMap, ) -> Self { + let store = ParquetStorage::new(Arc::clone(catalog_cache.object_store().object_store())); Self { catalog_cache, store, @@ -730,7 +730,6 @@ pub mod tests { catalog.object_store(), &Handle::current(), )), - ParquetStorage::new(catalog.object_store()), catalog.metric_registry(), HashMap::from([(parquet_file.id, load_settings)]), ); diff --git a/querier/src/database.rs b/querier/src/database.rs index f433e33201..a877e4838b 100644 --- a/querier/src/database.rs +++ b/querier/src/database.rs @@ -9,7 +9,6 @@ use backoff::{Backoff, BackoffConfig}; use data_types::{Namespace, ShardIndex}; use iox_catalog::interface::Catalog; use iox_query::exec::Executor; -use parquet_file::storage::ParquetStorage; use service_common::QueryDatabaseProvider; use sharder::JumpHash; use snafu::Snafu; @@ -107,7 +106,6 @@ impl QuerierDatabase { pub async fn new( catalog_cache: Arc, metric_registry: Arc, - store: ParquetStorage, exec: Arc, ingester_connection: Option>, max_concurrent_queries: usize, @@ -124,7 +122,6 @@ impl QuerierDatabase { let chunk_adapter = Arc::new(ChunkAdapter::new( Arc::clone(&catalog_cache), - store, Arc::clone(&metric_registry), Default::default(), )); @@ -263,7 +260,6 @@ mod tests { QuerierDatabase::new( catalog_cache, catalog.metric_registry(), - ParquetStorage::new(catalog.object_store()), catalog.exec(), Some(create_ingester_connection_for_testing()), QuerierDatabase::MAX_CONCURRENT_QUERIES_MAX.saturating_add(1), @@ -289,7 +285,6 @@ mod tests { QuerierDatabase::new( catalog_cache, catalog.metric_registry(), - ParquetStorage::new(catalog.object_store()), catalog.exec(), Some(create_ingester_connection_for_testing()), QuerierDatabase::MAX_CONCURRENT_QUERIES_MAX, @@ -316,7 +311,6 @@ mod tests { let db = QuerierDatabase::new( catalog_cache, catalog.metric_registry(), - ParquetStorage::new(catalog.object_store()), catalog.exec(), Some(create_ingester_connection_for_testing()), QuerierDatabase::MAX_CONCURRENT_QUERIES_MAX, @@ -347,7 +341,6 @@ mod tests { let db = QuerierDatabase::new( catalog_cache, catalog.metric_registry(), - ParquetStorage::new(catalog.object_store()), catalog.exec(), Some(create_ingester_connection_for_testing()), QuerierDatabase::MAX_CONCURRENT_QUERIES_MAX, diff --git a/querier/src/handler.rs b/querier/src/handler.rs index ea4ce7fe96..12aadd22f1 100644 --- a/querier/src/handler.rs +++ b/querier/src/handler.rs @@ -135,7 +135,6 @@ mod tests { use iox_query::exec::Executor; use iox_time::{MockProvider, Time}; use object_store::memory::InMemory; - use parquet_file::storage::ParquetStorage; use std::time::Duration; use tokio::runtime::Handle; @@ -191,7 +190,6 @@ mod tests { QuerierDatabase::new( catalog_cache, metric_registry, - ParquetStorage::new(object_store), exec, Some(create_ingester_connection_for_testing()), QuerierDatabase::MAX_CONCURRENT_QUERIES_MAX, diff --git a/querier/src/namespace/mod.rs b/querier/src/namespace/mod.rs index 1985716139..2c0b2a490a 100644 --- a/querier/src/namespace/mod.rs +++ b/querier/src/namespace/mod.rs @@ -10,7 +10,6 @@ use crate::{ }; use data_types::{NamespaceId, ParquetFileId, ShardIndex}; use iox_query::exec::Executor; -use parquet_file::storage::ParquetStorage; use sharder::JumpHash; use std::{collections::HashMap, sync::Arc}; @@ -101,7 +100,6 @@ impl QuerierNamespace { #[allow(clippy::too_many_arguments)] pub fn new_testing( catalog_cache: Arc, - store: ParquetStorage, metric_registry: Arc, name: Arc, ns: Arc, @@ -114,7 +112,6 @@ impl QuerierNamespace { let time_provider = catalog_cache.time_provider(); let chunk_adapter = Arc::new(ChunkAdapter::new( catalog_cache, - store, metric_registry, load_settings, )); diff --git a/querier/src/namespace/test_util.rs b/querier/src/namespace/test_util.rs index 29a13bf60f..a872b325fe 100644 --- a/querier/src/namespace/test_util.rs +++ b/querier/src/namespace/test_util.rs @@ -5,7 +5,6 @@ use crate::{ use data_types::{ShardIndex, TableId}; use iox_catalog::interface::get_schema_by_name; use iox_tests::util::TestNamespace; -use parquet_file::storage::ParquetStorage; use sharder::JumpHash; use std::sync::Arc; use tokio::runtime::Handle; @@ -38,7 +37,6 @@ pub async fn querier_namespace_with_limit( QuerierNamespace::new_testing( catalog_cache, - ParquetStorage::new(ns.catalog.object_store()), ns.catalog.metric_registry(), ns.namespace.name.clone().into(), cached_ns, diff --git a/querier/src/table/test_util.rs b/querier/src/table/test_util.rs index 3f75cad22f..961f9bde17 100644 --- a/querier/src/table/test_util.rs +++ b/querier/src/table/test_util.rs @@ -8,7 +8,6 @@ use data_types::{ChunkId, ParquetFileId, SequenceNumber, ShardIndex}; use iox_catalog::interface::get_schema_by_name; use iox_tests::util::{TestCatalog, TestPartition, TestShard, TestTable}; use mutable_batch_lp::test_helpers::lp_to_mutable_batch; -use parquet_file::storage::ParquetStorage; use schema::{selection::Selection, sort::SortKey, Schema}; use sharder::JumpHash; use std::{collections::HashMap, sync::Arc}; @@ -29,7 +28,6 @@ pub async fn querier_table( )); let chunk_adapter = Arc::new(ChunkAdapter::new( catalog_cache, - ParquetStorage::new(catalog.object_store()), catalog.metric_registry(), load_settings, )); diff --git a/query_tests/src/scenarios/util.rs b/query_tests/src/scenarios/util.rs index 14ef4ea294..453600423d 100644 --- a/query_tests/src/scenarios/util.rs +++ b/query_tests/src/scenarios/util.rs @@ -26,7 +26,6 @@ use iox_tests::util::{TestCatalog, TestNamespace, TestShard}; use itertools::Itertools; use mutable_batch_lp::LinesConverter; use once_cell::sync::Lazy; -use parquet_file::storage::ParquetStorage; use querier::{ IngesterConnectionImpl, IngesterFlightClient, IngesterFlightClientError, IngesterFlightClientQueryData, QuerierCatalogCache, QuerierChunkLoadSetting, QuerierNamespace, @@ -977,7 +976,6 @@ impl MockIngester { Arc::new(QuerierNamespace::new_testing( catalog_cache, - ParquetStorage::new(catalog.object_store()), catalog.metric_registry(), ns.namespace.name.clone().into(), Arc::new(schema.as_ref().into()),