ci: fix cargo deny (#4629)
* ci: fix cargo deny * chore: downgrade `socket2`, version 0.4.5 was yanked * chore: rename `query` to `iox_query` `query` is already taken on crates.io and yanked and I am getting tired of working around that.pull/24376/head
parent
7f60f69734
commit
52346642a0
|
|
@ -131,7 +131,7 @@ jobs:
|
|||
# https://github.com/EmbarkStudios/cargo-deny/issues/388
|
||||
name: Pull advisory-db
|
||||
command: |
|
||||
git clone --depth=1 https://github.com/RustSec/advisory-db.git /usr/local/cargo/advisory-dbs/github.com-dfdf83aac62a7e5e
|
||||
git clone --depth=1 https://github.com/RustSec/advisory-db.git /usr/local/cargo/advisory-dbs/github.com-2f857891b7f43c59
|
||||
- run:
|
||||
name: cargo-deny Checks
|
||||
command: cargo deny check -s --disable-fetch
|
||||
|
|
|
|||
|
|
@ -845,6 +845,7 @@ dependencies = [
|
|||
"datafusion 0.1.0",
|
||||
"futures",
|
||||
"iox_catalog",
|
||||
"iox_query",
|
||||
"iox_tests",
|
||||
"iox_time",
|
||||
"metric",
|
||||
|
|
@ -853,7 +854,6 @@ dependencies = [
|
|||
"parquet_file",
|
||||
"predicate",
|
||||
"querier",
|
||||
"query",
|
||||
"schema",
|
||||
"snafu",
|
||||
"test_helpers",
|
||||
|
|
@ -2181,6 +2181,7 @@ dependencies = [
|
|||
"influxdb_storage_client",
|
||||
"influxrpc_parser",
|
||||
"iox_catalog",
|
||||
"iox_query",
|
||||
"iox_time",
|
||||
"ioxd_common",
|
||||
"ioxd_compactor",
|
||||
|
|
@ -2200,7 +2201,6 @@ dependencies = [
|
|||
"parquet_file",
|
||||
"predicate",
|
||||
"predicates",
|
||||
"query",
|
||||
"rustyline",
|
||||
"schema",
|
||||
"serde_json",
|
||||
|
|
@ -2312,6 +2312,7 @@ dependencies = [
|
|||
"generated_types",
|
||||
"hyper",
|
||||
"iox_catalog",
|
||||
"iox_query",
|
||||
"iox_time",
|
||||
"lazy_static",
|
||||
"metric",
|
||||
|
|
@ -2325,7 +2326,6 @@ dependencies = [
|
|||
"pin-project",
|
||||
"predicate",
|
||||
"prost",
|
||||
"query",
|
||||
"schema",
|
||||
"snafu",
|
||||
"test_helpers",
|
||||
|
|
@ -2447,6 +2447,35 @@ dependencies = [
|
|||
"workspace-hack",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "iox_query"
|
||||
version = "0.1.0"
|
||||
dependencies = [
|
||||
"arrow",
|
||||
"arrow_util",
|
||||
"async-trait",
|
||||
"chrono",
|
||||
"croaring",
|
||||
"data_types",
|
||||
"datafusion 0.1.0",
|
||||
"datafusion_util",
|
||||
"executor",
|
||||
"futures",
|
||||
"hashbrown 0.12.1",
|
||||
"itertools",
|
||||
"observability_deps",
|
||||
"parking_lot 0.12.0",
|
||||
"predicate",
|
||||
"query_functions",
|
||||
"schema",
|
||||
"snafu",
|
||||
"test_helpers",
|
||||
"tokio",
|
||||
"tokio-stream",
|
||||
"trace",
|
||||
"workspace-hack",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "iox_tests"
|
||||
version = "0.1.0"
|
||||
|
|
@ -2456,12 +2485,12 @@ dependencies = [
|
|||
"data_types",
|
||||
"datafusion 0.1.0",
|
||||
"iox_catalog",
|
||||
"iox_query",
|
||||
"iox_time",
|
||||
"metric",
|
||||
"mutable_batch_lp",
|
||||
"object_store",
|
||||
"parquet_file",
|
||||
"query",
|
||||
"schema",
|
||||
"uuid 0.8.2",
|
||||
"workspace-hack",
|
||||
|
|
@ -2530,11 +2559,11 @@ dependencies = [
|
|||
"data_types",
|
||||
"hyper",
|
||||
"iox_catalog",
|
||||
"iox_query",
|
||||
"iox_time",
|
||||
"ioxd_common",
|
||||
"metric",
|
||||
"object_store",
|
||||
"query",
|
||||
"thiserror",
|
||||
"trace",
|
||||
"workspace-hack",
|
||||
|
|
@ -2550,10 +2579,10 @@ dependencies = [
|
|||
"hyper",
|
||||
"ingester",
|
||||
"iox_catalog",
|
||||
"iox_query",
|
||||
"ioxd_common",
|
||||
"metric",
|
||||
"object_store",
|
||||
"query",
|
||||
"thiserror",
|
||||
"trace",
|
||||
"workspace-hack",
|
||||
|
|
@ -2570,13 +2599,13 @@ dependencies = [
|
|||
"generated_types",
|
||||
"hyper",
|
||||
"iox_catalog",
|
||||
"iox_query",
|
||||
"iox_tests",
|
||||
"iox_time",
|
||||
"ioxd_common",
|
||||
"metric",
|
||||
"object_store",
|
||||
"querier",
|
||||
"query",
|
||||
"service_grpc_flight",
|
||||
"service_grpc_influxrpc",
|
||||
"tokio",
|
||||
|
|
@ -4132,6 +4161,7 @@ dependencies = [
|
|||
"generated_types",
|
||||
"influxdb_iox_client",
|
||||
"iox_catalog",
|
||||
"iox_query",
|
||||
"iox_tests",
|
||||
"iox_time",
|
||||
"metric",
|
||||
|
|
@ -4142,7 +4172,6 @@ dependencies = [
|
|||
"parquet_file",
|
||||
"pin-project",
|
||||
"predicate",
|
||||
"query",
|
||||
"rand",
|
||||
"schema",
|
||||
"service_common",
|
||||
|
|
@ -4158,35 +4187,6 @@ dependencies = [
|
|||
"workspace-hack",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "query"
|
||||
version = "0.1.0"
|
||||
dependencies = [
|
||||
"arrow",
|
||||
"arrow_util",
|
||||
"async-trait",
|
||||
"chrono",
|
||||
"croaring",
|
||||
"data_types",
|
||||
"datafusion 0.1.0",
|
||||
"datafusion_util",
|
||||
"executor",
|
||||
"futures",
|
||||
"hashbrown 0.12.1",
|
||||
"itertools",
|
||||
"observability_deps",
|
||||
"parking_lot 0.12.0",
|
||||
"predicate",
|
||||
"query_functions",
|
||||
"schema",
|
||||
"snafu",
|
||||
"test_helpers",
|
||||
"tokio",
|
||||
"tokio-stream",
|
||||
"trace",
|
||||
"workspace-hack",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "query_functions"
|
||||
version = "0.1.0"
|
||||
|
|
@ -4222,6 +4222,7 @@ dependencies = [
|
|||
"influxdb_iox_client",
|
||||
"ingester",
|
||||
"iox_catalog",
|
||||
"iox_query",
|
||||
"iox_tests",
|
||||
"itertools",
|
||||
"mutable_batch",
|
||||
|
|
@ -4230,7 +4231,6 @@ dependencies = [
|
|||
"predicate",
|
||||
"pretty_assertions",
|
||||
"querier",
|
||||
"query",
|
||||
"schema",
|
||||
"snafu",
|
||||
"tempfile",
|
||||
|
|
@ -5033,8 +5033,8 @@ version = "0.1.0"
|
|||
dependencies = [
|
||||
"async-trait",
|
||||
"datafusion 0.1.0",
|
||||
"iox_query",
|
||||
"predicate",
|
||||
"query",
|
||||
"workspace-hack",
|
||||
]
|
||||
|
||||
|
|
@ -5064,10 +5064,10 @@ dependencies = [
|
|||
"datafusion 0.1.0",
|
||||
"futures",
|
||||
"generated_types",
|
||||
"iox_query",
|
||||
"observability_deps",
|
||||
"pin-project",
|
||||
"prost",
|
||||
"query",
|
||||
"serde",
|
||||
"serde_json",
|
||||
"service_common",
|
||||
|
|
@ -5088,13 +5088,13 @@ dependencies = [
|
|||
"futures",
|
||||
"generated_types",
|
||||
"influxdb_storage_client",
|
||||
"iox_query",
|
||||
"metric",
|
||||
"observability_deps",
|
||||
"panic_logging",
|
||||
"parking_lot 0.12.0",
|
||||
"predicate",
|
||||
"prost",
|
||||
"query",
|
||||
"query_functions",
|
||||
"regex",
|
||||
"schema",
|
||||
|
|
@ -5291,12 +5291,12 @@ checksum = "45456094d1983e2ee2a18fdfebce3189fa451699d0502cb8e3b49dba5ba41451"
|
|||
|
||||
[[package]]
|
||||
name = "socket2"
|
||||
version = "0.4.5"
|
||||
version = "0.4.4"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "ca642ba17f8b2995138b1d7711829c92e98c0a25ea019de790f4f09279c4e296"
|
||||
checksum = "66d72b759436ae32898a2af0a14218dbf55efde3feeb170eb623637db85ee1e0"
|
||||
dependencies = [
|
||||
"libc",
|
||||
"windows-sys 0.36.1",
|
||||
"winapi",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
|
|
|
|||
|
|
@ -26,6 +26,7 @@ members = [
|
|||
"iox_catalog",
|
||||
"iox_data_generator",
|
||||
"iox_gitops_adapter",
|
||||
"iox_query",
|
||||
"iox_tests",
|
||||
"iox_time",
|
||||
"ioxd_common",
|
||||
|
|
@ -48,7 +49,6 @@ members = [
|
|||
"parquet_file",
|
||||
"predicate",
|
||||
"querier",
|
||||
"query",
|
||||
"query_functions",
|
||||
"query_tests",
|
||||
"read_buffer",
|
||||
|
|
|
|||
|
|
@ -18,7 +18,7 @@ object_store = "0.0.1"
|
|||
observability_deps = { path = "../observability_deps" }
|
||||
parquet_file = { path = "../parquet_file" }
|
||||
predicate = { path = "../predicate" }
|
||||
query = { path = "../query" }
|
||||
iox_query = { path = "../iox_query" }
|
||||
schema = { path = "../schema" }
|
||||
snafu = "0.7"
|
||||
thiserror = "1.0"
|
||||
|
|
|
|||
|
|
@ -17,6 +17,13 @@ use data_types::{
|
|||
};
|
||||
use datafusion::error::DataFusionError;
|
||||
use iox_catalog::interface::{Catalog, Transaction, INITIAL_COMPACTION_LEVEL};
|
||||
use iox_query::{
|
||||
exec::{Executor, ExecutorType},
|
||||
frontend::reorg::ReorgPlanner,
|
||||
provider::overlap::group_potential_duplicates,
|
||||
util::compute_timenanosecond_min_max,
|
||||
QueryChunk,
|
||||
};
|
||||
use iox_time::{Time, TimeProvider};
|
||||
use metric::{Attributes, Metric, U64Counter, U64Gauge, U64Histogram, U64HistogramOptions};
|
||||
use object_store::DynObjectStore;
|
||||
|
|
@ -25,13 +32,6 @@ use parquet_file::{
|
|||
metadata::{IoxMetadata, IoxParquetMetaData},
|
||||
ParquetFilePath,
|
||||
};
|
||||
use query::{
|
||||
exec::{Executor, ExecutorType},
|
||||
frontend::reorg::ReorgPlanner,
|
||||
provider::overlap::group_potential_duplicates,
|
||||
util::compute_timenanosecond_min_max,
|
||||
QueryChunk,
|
||||
};
|
||||
use schema::sort::SortKey;
|
||||
use snafu::{ensure, OptionExt, ResultExt, Snafu};
|
||||
use std::{
|
||||
|
|
@ -87,7 +87,7 @@ pub enum Error {
|
|||
|
||||
#[snafu(display("Error building compact logical plan {}", source))]
|
||||
CompactLogicalPlan {
|
||||
source: query::frontend::reorg::Error,
|
||||
source: iox_query::frontend::reorg::Error,
|
||||
},
|
||||
|
||||
#[snafu(display("Error building compact physical plan {}", source))]
|
||||
|
|
@ -103,7 +103,7 @@ pub enum Error {
|
|||
RowCountTypeConversion { source: std::num::TryFromIntError },
|
||||
|
||||
#[snafu(display("Error computing min and max for record batches: {}", source))]
|
||||
MinMax { source: query::util::Error },
|
||||
MinMax { source: iox_query::util::Error },
|
||||
|
||||
#[snafu(display("Error while starting catalog transaction {}", source))]
|
||||
Transaction {
|
||||
|
|
@ -1108,13 +1108,13 @@ mod tests {
|
|||
use data_types::{ChunkId, KafkaPartition, NamespaceId, ParquetFileParams, SequenceNumber};
|
||||
use futures::TryStreamExt;
|
||||
use iox_catalog::interface::INITIAL_COMPACTION_LEVEL;
|
||||
use iox_query::test::{raw_data, TestChunk};
|
||||
use iox_tests::util::TestCatalog;
|
||||
use iox_time::SystemProvider;
|
||||
use querier::{
|
||||
cache::CatalogCache,
|
||||
chunk::{collect_read_filter, ParquetChunkAdapter},
|
||||
};
|
||||
use query::test::{raw_data, TestChunk};
|
||||
use std::sync::atomic::{AtomicI64, Ordering};
|
||||
|
||||
// Simulate unique ID generation
|
||||
|
|
|
|||
|
|
@ -8,10 +8,10 @@ use futures::{
|
|||
select, FutureExt, TryFutureExt,
|
||||
};
|
||||
use iox_catalog::interface::Catalog;
|
||||
use iox_query::exec::Executor;
|
||||
use iox_time::TimeProvider;
|
||||
use object_store::DynObjectStore;
|
||||
use observability_deps::tracing::*;
|
||||
use query::exec::Executor;
|
||||
use std::sync::Arc;
|
||||
use thiserror::Error;
|
||||
use tokio::task::{JoinError, JoinHandle};
|
||||
|
|
|
|||
|
|
@ -5,13 +5,13 @@ use data_types::{
|
|||
TimestampMinMax, Tombstone,
|
||||
};
|
||||
use datafusion::physical_plan::SendableRecordBatchStream;
|
||||
use observability_deps::tracing::trace;
|
||||
use parquet_file::chunk::ParquetChunk;
|
||||
use predicate::{delete_predicate::tombstones_to_delete_predicates, Predicate, PredicateMatch};
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::{stringset::StringSet, IOxSessionContext},
|
||||
QueryChunk, QueryChunkError, QueryChunkMeta,
|
||||
};
|
||||
use observability_deps::tracing::trace;
|
||||
use parquet_file::chunk::ParquetChunk;
|
||||
use predicate::{delete_predicate::tombstones_to_delete_predicates, Predicate, PredicateMatch};
|
||||
use schema::{merge::SchemaMerger, selection::Selection, sort::SortKey, Schema};
|
||||
use snafu::{ResultExt, Snafu};
|
||||
use std::sync::Arc;
|
||||
|
|
|
|||
|
|
@ -26,6 +26,8 @@ ignore = [
|
|||
# why needed: used by `criterion`
|
||||
# upstream issue: https://github.com/bheisler/criterion.rs/issues/534
|
||||
"RUSTSEC-2021-0127",
|
||||
|
||||
"query",
|
||||
]
|
||||
|
||||
[licenses]
|
||||
|
|
|
|||
|
|
@ -43,5 +43,5 @@ We hold monthly Tech Talks that explain the project's technical underpinnings. Y
|
|||
* [Data Deduplication & Sort Key](dedup_and_sort.md)
|
||||
* Transactions and locks (to be written & linked)
|
||||
* [Thoughts on using multiple cores / thread pools](multi_core_tasks.md)
|
||||
* [Query Engine Docs](../query/README.md)
|
||||
* [Query Engine Docs](../iox_query/README.md)
|
||||
* [Notes on the use of local filesystems](local_filesystems.md)
|
||||
|
|
|
|||
|
|
@ -27,7 +27,7 @@ object_store_metrics = { path = "../object_store_metrics" }
|
|||
observability_deps = { path = "../observability_deps" }
|
||||
panic_logging = { path = "../panic_logging" }
|
||||
parquet_file = { path = "../parquet_file" }
|
||||
query = { path = "../query" }
|
||||
iox_query = { path = "../iox_query" }
|
||||
schema = { path = "../schema" }
|
||||
iox_time = { path = "../iox_time" }
|
||||
trace_exporters = { path = "../trace_exporters" }
|
||||
|
|
|
|||
|
|
@ -8,6 +8,7 @@ use clap_blocks::{
|
|||
object_store::ObjectStoreConfig, run_config::RunConfig, socket_addr::SocketAddr,
|
||||
write_buffer::WriteBufferConfig,
|
||||
};
|
||||
use iox_query::exec::Executor;
|
||||
use iox_time::{SystemProvider, TimeProvider};
|
||||
use ioxd_common::{
|
||||
server_type::{CommonServerState, CommonServerStateError},
|
||||
|
|
@ -19,7 +20,6 @@ use ioxd_querier::{create_querier_server_type, QuerierServerTypeArgs};
|
|||
use ioxd_router::create_router_server_type;
|
||||
use object_store::DynObjectStore;
|
||||
use observability_deps::tracing::*;
|
||||
use query::exec::Executor;
|
||||
use std::{path::PathBuf, sync::Arc};
|
||||
use thiserror::Error;
|
||||
use trace_exporters::TracingConfig;
|
||||
|
|
|
|||
|
|
@ -1,10 +1,10 @@
|
|||
//! Implementation of command line option for running the compactor
|
||||
|
||||
use iox_query::exec::Executor;
|
||||
use iox_time::{SystemProvider, TimeProvider};
|
||||
use object_store::DynObjectStore;
|
||||
use object_store_metrics::ObjectStoreMetrics;
|
||||
use observability_deps::tracing::*;
|
||||
use query::exec::Executor;
|
||||
use std::sync::Arc;
|
||||
use thiserror::Error;
|
||||
|
||||
|
|
|
|||
|
|
@ -5,6 +5,7 @@ use clap_blocks::{
|
|||
catalog_dsn::CatalogDsnConfig, ingester::IngesterConfig, run_config::RunConfig,
|
||||
write_buffer::WriteBufferConfig,
|
||||
};
|
||||
use iox_query::exec::Executor;
|
||||
use iox_time::{SystemProvider, TimeProvider};
|
||||
use ioxd_common::server_type::{CommonServerState, CommonServerStateError};
|
||||
use ioxd_common::Service;
|
||||
|
|
@ -12,7 +13,6 @@ use ioxd_ingester::create_ingester_server_type;
|
|||
use object_store::DynObjectStore;
|
||||
use object_store_metrics::ObjectStoreMetrics;
|
||||
use observability_deps::tracing::*;
|
||||
use query::exec::Executor;
|
||||
use std::sync::Arc;
|
||||
use thiserror::Error;
|
||||
|
||||
|
|
|
|||
|
|
@ -1,11 +1,11 @@
|
|||
//! Implementation of command line option for running the querier
|
||||
|
||||
use clap_blocks::querier::QuerierConfig;
|
||||
use iox_query::exec::Executor;
|
||||
use iox_time::{SystemProvider, TimeProvider};
|
||||
use object_store::DynObjectStore;
|
||||
use object_store_metrics::ObjectStoreMetrics;
|
||||
use observability_deps::tracing::*;
|
||||
use query::exec::Executor;
|
||||
use std::sync::Arc;
|
||||
use thiserror::Error;
|
||||
|
||||
|
|
|
|||
|
|
@ -30,7 +30,7 @@ parquet_file = { path = "../parquet_file" }
|
|||
pin-project = "1.0"
|
||||
predicate = { path = "../predicate" }
|
||||
prost = "0.10"
|
||||
query = { path = "../query" }
|
||||
iox_query = { path = "../iox_query" }
|
||||
schema = { path = "../schema" }
|
||||
snafu = "0.7"
|
||||
thiserror = "1.0"
|
||||
|
|
|
|||
|
|
@ -5,14 +5,14 @@ use arrow::record_batch::RecordBatch;
|
|||
use data_types::{NamespaceId, PartitionInfo};
|
||||
use datafusion::{error::DataFusionError, physical_plan::SendableRecordBatchStream};
|
||||
use iox_catalog::interface::INITIAL_COMPACTION_LEVEL;
|
||||
use iox_time::{Time, TimeProvider};
|
||||
use parquet_file::metadata::IoxMetadata;
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::{Executor, ExecutorType},
|
||||
frontend::reorg::ReorgPlanner,
|
||||
util::compute_timenanosecond_min_max,
|
||||
QueryChunk, QueryChunkMeta,
|
||||
};
|
||||
use iox_time::{Time, TimeProvider};
|
||||
use parquet_file::metadata::IoxMetadata;
|
||||
use schema::sort::{adjust_sort_key_columns, compute_sort_key, SortKey};
|
||||
use snafu::{ResultExt, Snafu};
|
||||
use std::sync::Arc;
|
||||
|
|
@ -22,7 +22,7 @@ use std::sync::Arc;
|
|||
pub enum Error {
|
||||
#[snafu(display("Error while building logical plan for Ingester's compaction"))]
|
||||
LogicalPlan {
|
||||
source: query::frontend::reorg::Error,
|
||||
source: iox_query::frontend::reorg::Error,
|
||||
},
|
||||
|
||||
#[snafu(display("Error while building physical plan for Ingester's compaction"))]
|
||||
|
|
@ -46,7 +46,7 @@ pub enum Error {
|
|||
RowCountTypeConversion { source: std::num::TryFromIntError },
|
||||
|
||||
#[snafu(display("Error computing min and max for record batches: {}", source))]
|
||||
MinMax { source: query::util::Error },
|
||||
MinMax { source: iox_query::util::Error },
|
||||
}
|
||||
|
||||
/// A specialized `Error` for Ingester's Compact errors
|
||||
|
|
|
|||
|
|
@ -17,6 +17,7 @@ use data_types::{
|
|||
use datafusion::physical_plan::SendableRecordBatchStream;
|
||||
use dml::DmlOperation;
|
||||
use iox_catalog::interface::Catalog;
|
||||
use iox_query::exec::Executor;
|
||||
use iox_time::SystemProvider;
|
||||
use metric::U64Counter;
|
||||
use mutable_batch::MutableBatch;
|
||||
|
|
@ -24,7 +25,6 @@ use object_store::DynObjectStore;
|
|||
use observability_deps::tracing::{debug, warn};
|
||||
use parking_lot::RwLock;
|
||||
use predicate::Predicate;
|
||||
use query::exec::Executor;
|
||||
use schema::{selection::Selection, Schema};
|
||||
use snafu::{OptionExt, ResultExt, Snafu};
|
||||
use std::{
|
||||
|
|
|
|||
|
|
@ -21,11 +21,11 @@ use futures::{
|
|||
};
|
||||
use generated_types::ingester::IngesterQueryRequest;
|
||||
use iox_catalog::interface::Catalog;
|
||||
use iox_query::exec::Executor;
|
||||
use iox_time::{SystemProvider, TimeProvider};
|
||||
use metric::{Metric, U64Histogram, U64HistogramOptions};
|
||||
use object_store::DynObjectStore;
|
||||
use observability_deps::tracing::*;
|
||||
use query::exec::Executor;
|
||||
use snafu::{ResultExt, Snafu};
|
||||
use std::{collections::BTreeMap, sync::Arc, time::Duration};
|
||||
use tokio::task::{JoinError, JoinHandle};
|
||||
|
|
|
|||
|
|
@ -84,9 +84,9 @@ mod tests {
|
|||
use data_types::{NamespaceId, PartitionId, SequenceNumber, SequencerId, TableId};
|
||||
use futures::{StreamExt, TryStreamExt};
|
||||
use iox_catalog::interface::INITIAL_COMPACTION_LEVEL;
|
||||
use iox_query::test::{raw_data, TestChunk};
|
||||
use iox_time::Time;
|
||||
use object_store::memory::InMemory;
|
||||
use query::test::{raw_data, TestChunk};
|
||||
use std::sync::Arc;
|
||||
use uuid::Uuid;
|
||||
|
||||
|
|
|
|||
|
|
@ -14,13 +14,13 @@ use datafusion::{
|
|||
},
|
||||
};
|
||||
use generated_types::ingester::IngesterQueryRequest;
|
||||
use observability_deps::tracing::debug;
|
||||
use predicate::Predicate;
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::{Executor, ExecutorType},
|
||||
frontend::reorg::ReorgPlanner,
|
||||
QueryChunkMeta,
|
||||
};
|
||||
use observability_deps::tracing::debug;
|
||||
use predicate::Predicate;
|
||||
use schema::{
|
||||
merge::{merge_record_batch_schemas, SchemaMerger},
|
||||
selection::Selection,
|
||||
|
|
@ -37,7 +37,7 @@ pub enum Error {
|
|||
|
||||
#[snafu(display("Error building logical plan for querying Ingester data to send to Querier"))]
|
||||
LogicalPlan {
|
||||
source: query::frontend::reorg::Error,
|
||||
source: iox_query::frontend::reorg::Error,
|
||||
},
|
||||
|
||||
#[snafu(display(
|
||||
|
|
|
|||
|
|
@ -16,16 +16,16 @@ use datafusion::{
|
|||
},
|
||||
};
|
||||
use datafusion_util::batch_filter;
|
||||
use iox_query::{
|
||||
exec::{stringset::StringSet, IOxSessionContext},
|
||||
util::{df_physical_expr_from_schema_and_expr, MissingColumnsToNull},
|
||||
QueryChunk, QueryChunkError, QueryChunkMeta,
|
||||
};
|
||||
use observability_deps::tracing::{debug, trace};
|
||||
use predicate::{
|
||||
delete_predicate::{tombstones_to_delete_predicates, tombstones_to_delete_predicates_iter},
|
||||
Predicate, PredicateMatch,
|
||||
};
|
||||
use query::{
|
||||
exec::{stringset::StringSet, IOxSessionContext},
|
||||
util::{df_physical_expr_from_schema_and_expr, MissingColumnsToNull},
|
||||
QueryChunk, QueryChunkError, QueryChunkMeta,
|
||||
};
|
||||
use schema::{merge::merge_record_batch_schemas, selection::Selection, sort::SortKey, Schema};
|
||||
use snafu::{ResultExt, Snafu};
|
||||
use std::sync::Arc;
|
||||
|
|
|
|||
|
|
@ -20,10 +20,10 @@ use iox_catalog::{
|
|||
interface::{Catalog, INITIAL_COMPACTION_LEVEL},
|
||||
mem::MemCatalog,
|
||||
};
|
||||
use iox_query::test::{raw_data, TestChunk};
|
||||
use iox_time::{SystemProvider, Time, TimeProvider};
|
||||
use object_store::memory::InMemory;
|
||||
use parquet_file::metadata::IoxMetadata;
|
||||
use query::test::{raw_data, TestChunk};
|
||||
use schema::sort::SortKey;
|
||||
use std::{collections::BTreeMap, sync::Arc};
|
||||
use uuid::Uuid;
|
||||
|
|
@ -677,7 +677,7 @@ pub fn make_ingester_data(two_partitions: bool, loc: DataLocation) -> IngesterDa
|
|||
let metrics: Arc<metric::Registry> = Default::default();
|
||||
let catalog: Arc<dyn Catalog> = Arc::new(MemCatalog::new(Arc::clone(&metrics)));
|
||||
let object_store = Arc::new(InMemory::new());
|
||||
let exec = Arc::new(query::exec::Executor::new(1));
|
||||
let exec = Arc::new(iox_query::exec::Executor::new(1));
|
||||
|
||||
// Make data for one sequencer/shard and two tables
|
||||
let seq_id = SequencerId::new(1);
|
||||
|
|
@ -730,7 +730,7 @@ pub async fn make_ingester_data_with_tombstones(loc: DataLocation) -> IngesterDa
|
|||
let metrics: Arc<metric::Registry> = Default::default();
|
||||
let catalog: Arc<dyn Catalog> = Arc::new(MemCatalog::new(metrics));
|
||||
let object_store = Arc::new(InMemory::new());
|
||||
let exec = Arc::new(query::exec::Executor::new(1));
|
||||
let exec = Arc::new(iox_query::exec::Executor::new(1));
|
||||
|
||||
// Make data for one sequencer/shard and two tables
|
||||
let seq_id = SequencerId::new(1);
|
||||
|
|
@ -845,7 +845,7 @@ pub(crate) fn make_partitions(
|
|||
|
||||
/// Make data for one partition with tombstones
|
||||
pub(crate) async fn make_one_partition_with_tombstones(
|
||||
exec: &query::exec::Executor,
|
||||
exec: &iox_query::exec::Executor,
|
||||
loc: DataLocation,
|
||||
sequencer_id: SequencerId,
|
||||
table_id: TableId,
|
||||
|
|
|
|||
|
|
@ -1,5 +1,5 @@
|
|||
[package]
|
||||
name = "query"
|
||||
name = "iox_query"
|
||||
version = "0.1.0"
|
||||
authors = ["Andrew Lamb <andrew@nerdnetworks.org>"]
|
||||
edition = "2021"
|
||||
|
|
@ -383,7 +383,7 @@ pub fn compute_sort_key<'a>(summaries: impl Iterator<Item = &'a TableSummary>) -
|
|||
// error[E0433]: failed to resolve: could not find `test` in `storage`
|
||||
// --> src/server/mutable_buffer_routes.rs:353:19
|
||||
// |
|
||||
// 353 | use query::test::TestDatabaseStore;
|
||||
// 353 | use iox_query::test::TestDatabaseStore;
|
||||
// | ^^^^ could not find `test` in `query`
|
||||
|
||||
//
|
||||
|
|
@ -16,7 +16,7 @@ metric = { path = "../metric" }
|
|||
mutable_batch_lp = { path = "../mutable_batch_lp" }
|
||||
object_store = "0.0.1"
|
||||
parquet_file = { path = "../parquet_file" }
|
||||
query = { path = "../query" }
|
||||
iox_query = { path = "../iox_query" }
|
||||
schema = { path = "../schema" }
|
||||
uuid = { version = "0.8", features = ["v4"] }
|
||||
workspace-hack = { path = "../workspace-hack"}
|
||||
|
|
|
|||
|
|
@ -15,6 +15,7 @@ use iox_catalog::{
|
|||
interface::{Catalog, PartitionRepo, INITIAL_COMPACTION_LEVEL},
|
||||
mem::MemCatalog,
|
||||
};
|
||||
use iox_query::{exec::Executor, provider::RecordBatchDeduplicator, util::arrow_sort_key_exprs};
|
||||
use iox_time::{MockProvider, Time, TimeProvider};
|
||||
use mutable_batch_lp::test_helpers::lp_to_mutable_batch;
|
||||
use object_store::{memory::InMemory, DynObjectStore};
|
||||
|
|
@ -22,7 +23,6 @@ use parquet_file::{
|
|||
metadata::{IoxMetadata, IoxParquetMetaData},
|
||||
ParquetFilePath,
|
||||
};
|
||||
use query::{exec::Executor, provider::RecordBatchDeduplicator, util::arrow_sort_key_exprs};
|
||||
use schema::{
|
||||
selection::Selection,
|
||||
sort::{adjust_sort_key_columns, SortKey, SortKeyBuilder},
|
||||
|
|
|
|||
|
|
@ -13,7 +13,7 @@ data_types = { path = "../data_types" }
|
|||
iox_catalog = { path = "../iox_catalog" }
|
||||
ioxd_common = { path = "../ioxd_common" }
|
||||
metric = { path = "../metric" }
|
||||
query = { path = "../query" }
|
||||
iox_query = { path = "../iox_query" }
|
||||
object_store = "0.0.1"
|
||||
iox_time = { path = "../iox_time" }
|
||||
trace = { path = "../trace" }
|
||||
|
|
|
|||
|
|
@ -7,6 +7,7 @@ use compactor::{
|
|||
use data_types::KafkaPartition;
|
||||
use hyper::{Body, Request, Response};
|
||||
use iox_catalog::interface::Catalog;
|
||||
use iox_query::exec::Executor;
|
||||
use iox_time::TimeProvider;
|
||||
use ioxd_common::{
|
||||
add_service,
|
||||
|
|
@ -18,7 +19,6 @@ use ioxd_common::{
|
|||
};
|
||||
use metric::Registry;
|
||||
use object_store::DynObjectStore;
|
||||
use query::exec::Executor;
|
||||
use std::{
|
||||
fmt::{Debug, Display},
|
||||
sync::Arc,
|
||||
|
|
|
|||
|
|
@ -12,7 +12,7 @@ iox_catalog = { path = "../iox_catalog" }
|
|||
ioxd_common = { path = "../ioxd_common" }
|
||||
metric = { path = "../metric" }
|
||||
object_store = "0.0.1"
|
||||
query = { path = "../query" }
|
||||
iox_query = { path = "../iox_query" }
|
||||
trace = { path = "../trace" }
|
||||
write_buffer = { path = "../write_buffer" }
|
||||
|
||||
|
|
|
|||
|
|
@ -8,6 +8,7 @@ use ingester::{
|
|||
server::{grpc::GrpcDelegate, http::HttpDelegate, IngesterServer},
|
||||
};
|
||||
use iox_catalog::interface::Catalog;
|
||||
use iox_query::exec::Executor;
|
||||
use ioxd_common::{
|
||||
add_service,
|
||||
http::error::{HttpApiError, HttpApiErrorCode, HttpApiErrorSource},
|
||||
|
|
@ -18,7 +19,6 @@ use ioxd_common::{
|
|||
};
|
||||
use metric::Registry;
|
||||
use object_store::DynObjectStore;
|
||||
use query::exec::Executor;
|
||||
use std::{
|
||||
collections::BTreeMap,
|
||||
fmt::{Debug, Display},
|
||||
|
|
|
|||
|
|
@ -14,7 +14,7 @@ ioxd_common = { path = "../ioxd_common" }
|
|||
metric = { path = "../metric" }
|
||||
object_store = "0.0.1"
|
||||
querier = { path = "../querier" }
|
||||
query = { path = "../query" }
|
||||
iox_query = { path = "../iox_query" }
|
||||
service_grpc_flight = { path = "../service_grpc_flight" }
|
||||
service_grpc_influxrpc = { path = "../service_grpc_influxrpc" }
|
||||
iox_time = { path = "../iox_time" }
|
||||
|
|
|
|||
|
|
@ -6,6 +6,7 @@ use std::{
|
|||
use async_trait::async_trait;
|
||||
use hyper::{Body, Request, Response};
|
||||
use iox_catalog::interface::Catalog;
|
||||
use iox_query::exec::Executor;
|
||||
use iox_time::TimeProvider;
|
||||
use metric::Registry;
|
||||
use object_store::DynObjectStore;
|
||||
|
|
@ -13,7 +14,6 @@ use querier::{
|
|||
create_ingester_connection, QuerierCatalogCache, QuerierDatabase, QuerierHandler,
|
||||
QuerierHandlerImpl, QuerierServer,
|
||||
};
|
||||
use query::exec::Executor;
|
||||
use trace::TraceCollector;
|
||||
|
||||
use ioxd_common::{
|
||||
|
|
|
|||
|
|
@ -23,7 +23,7 @@ parking_lot = "0.12"
|
|||
parquet_file = { path = "../parquet_file" }
|
||||
pin-project = "1.0"
|
||||
predicate = { path = "../predicate" }
|
||||
query = { path = "../query" }
|
||||
iox_query = { path = "../iox_query" }
|
||||
rand = "0.8.3"
|
||||
service_common = { path = "../service_common" }
|
||||
service_grpc_schema = { path = "../service_grpc_schema" }
|
||||
|
|
|
|||
|
|
@ -8,12 +8,12 @@ use data_types::{
|
|||
};
|
||||
use futures::StreamExt;
|
||||
use iox_catalog::interface::Catalog;
|
||||
use iox_query::{exec::IOxSessionContext, QueryChunk};
|
||||
use iox_time::TimeProvider;
|
||||
use object_store::DynObjectStore;
|
||||
use parquet_file::chunk::{
|
||||
new_parquet_chunk, ChunkMetrics as ParquetChunkMetrics, DecodedParquetFile, ParquetChunk,
|
||||
};
|
||||
use query::{exec::IOxSessionContext, QueryChunk};
|
||||
use schema::{selection::Selection, sort::SortKey};
|
||||
use std::sync::Arc;
|
||||
use uuid::Uuid;
|
||||
|
|
@ -292,8 +292,8 @@ pub mod tests {
|
|||
use super::*;
|
||||
use arrow::datatypes::DataType;
|
||||
use arrow_util::assert_batches_eq;
|
||||
use iox_query::QueryChunkMeta;
|
||||
use iox_tests::util::TestCatalog;
|
||||
use query::QueryChunkMeta;
|
||||
use schema::{builder::SchemaBuilder, sort::SortKeyBuilder};
|
||||
|
||||
#[tokio::test]
|
||||
|
|
|
|||
|
|
@ -2,9 +2,9 @@ use crate::chunk::{ChunkStorage, QuerierChunk};
|
|||
use data_types::{
|
||||
ChunkId, ChunkOrder, DeletePredicate, PartitionId, TableSummary, TimestampMinMax,
|
||||
};
|
||||
use iox_query::{QueryChunk, QueryChunkError, QueryChunkMeta};
|
||||
use observability_deps::tracing::debug;
|
||||
use predicate::PredicateMatch;
|
||||
use query::{QueryChunk, QueryChunkError, QueryChunkMeta};
|
||||
use schema::{sort::SortKey, Schema};
|
||||
use snafu::{ResultExt, Snafu};
|
||||
use std::sync::Arc;
|
||||
|
|
@ -86,10 +86,10 @@ impl QueryChunk for QuerierChunk {
|
|||
|
||||
fn column_names(
|
||||
&self,
|
||||
_ctx: query::exec::IOxSessionContext,
|
||||
_ctx: iox_query::exec::IOxSessionContext,
|
||||
predicate: &predicate::Predicate,
|
||||
columns: schema::selection::Selection<'_>,
|
||||
) -> Result<Option<query::exec::stringset::StringSet>, QueryChunkError> {
|
||||
) -> Result<Option<iox_query::exec::stringset::StringSet>, QueryChunkError> {
|
||||
match &self.storage {
|
||||
ChunkStorage::Parquet { chunk, .. } => {
|
||||
if !predicate.is_empty() {
|
||||
|
|
@ -103,10 +103,10 @@ impl QueryChunk for QuerierChunk {
|
|||
|
||||
fn column_values(
|
||||
&self,
|
||||
_ctx: query::exec::IOxSessionContext,
|
||||
_ctx: iox_query::exec::IOxSessionContext,
|
||||
_column_name: &str,
|
||||
_predicate: &predicate::Predicate,
|
||||
) -> Result<Option<query::exec::stringset::StringSet>, QueryChunkError> {
|
||||
) -> Result<Option<iox_query::exec::stringset::StringSet>, QueryChunkError> {
|
||||
match &self.storage {
|
||||
ChunkStorage::Parquet { .. } => {
|
||||
// Since DataFusion can read Parquet, there is no advantage to
|
||||
|
|
@ -118,7 +118,7 @@ impl QueryChunk for QuerierChunk {
|
|||
|
||||
fn read_filter(
|
||||
&self,
|
||||
mut ctx: query::exec::IOxSessionContext,
|
||||
mut ctx: iox_query::exec::IOxSessionContext,
|
||||
predicate: &predicate::Predicate,
|
||||
selection: schema::selection::Selection<'_>,
|
||||
) -> Result<datafusion::physical_plan::SendableRecordBatchStream, QueryChunkError> {
|
||||
|
|
|
|||
|
|
@ -7,9 +7,9 @@ use crate::{
|
|||
use async_trait::async_trait;
|
||||
use backoff::{Backoff, BackoffConfig};
|
||||
use data_types::Namespace;
|
||||
use iox_query::exec::Executor;
|
||||
use object_store::DynObjectStore;
|
||||
use parking_lot::RwLock;
|
||||
use query::exec::Executor;
|
||||
use service_common::QueryDatabaseProvider;
|
||||
use std::{collections::HashMap, sync::Arc};
|
||||
|
||||
|
|
|
|||
|
|
@ -131,9 +131,9 @@ mod tests {
|
|||
use std::time::Duration;
|
||||
|
||||
use iox_catalog::mem::MemCatalog;
|
||||
use iox_query::exec::Executor;
|
||||
use iox_time::{MockProvider, Time};
|
||||
use object_store::memory::InMemory;
|
||||
use query::exec::Executor;
|
||||
|
||||
use crate::{cache::CatalogCache, create_ingester_connection_for_testing};
|
||||
|
||||
|
|
|
|||
|
|
@ -16,13 +16,13 @@ use generated_types::{
|
|||
influxdata::iox::ingester::v1::GetWriteInfoResponse, ingester::IngesterQueryRequest,
|
||||
write_info::merge_responses,
|
||||
};
|
||||
use observability_deps::tracing::{debug, trace};
|
||||
use predicate::{Predicate, PredicateMatch};
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::{stringset::StringSet, IOxSessionContext},
|
||||
util::compute_timenanosecond_min_max,
|
||||
QueryChunk, QueryChunkError, QueryChunkMeta,
|
||||
};
|
||||
use observability_deps::tracing::{debug, trace};
|
||||
use predicate::{Predicate, PredicateMatch};
|
||||
use schema::{selection::Selection, sort::SortKey, InfluxColumnType, InfluxFieldType, Schema};
|
||||
use snafu::{ensure, OptionExt, ResultExt, Snafu};
|
||||
use std::{any::Any, collections::HashMap, sync::Arc};
|
||||
|
|
|
|||
|
|
@ -6,8 +6,8 @@ use crate::{
|
|||
};
|
||||
use backoff::BackoffConfig;
|
||||
use data_types::{NamespaceId, NamespaceSchema};
|
||||
use iox_query::exec::Executor;
|
||||
use object_store::DynObjectStore;
|
||||
use query::exec::Executor;
|
||||
use schema::Schema;
|
||||
use std::{collections::HashMap, sync::Arc};
|
||||
|
||||
|
|
|
|||
|
|
@ -1,4 +1,4 @@
|
|||
//! This module contains implementations of [query] interfaces for [QuerierNamespace].
|
||||
//! This module contains implementations of [`iox_query`] interfaces for [QuerierNamespace].
|
||||
|
||||
use crate::{
|
||||
namespace::QuerierNamespace,
|
||||
|
|
@ -12,12 +12,12 @@ use datafusion::{
|
|||
catalog::{catalog::CatalogProvider, schema::SchemaProvider},
|
||||
datasource::TableProvider,
|
||||
};
|
||||
use observability_deps::tracing::{debug, trace};
|
||||
use predicate::{rpc_predicate::QueryDatabaseMeta, Predicate};
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::{ExecutionContextProvider, ExecutorType, IOxSessionContext},
|
||||
QueryChunk, QueryCompletedToken, QueryDatabase, QueryDatabaseError, QueryText, DEFAULT_SCHEMA,
|
||||
};
|
||||
use observability_deps::tracing::{debug, trace};
|
||||
use predicate::{rpc_predicate::QueryDatabaseMeta, Predicate};
|
||||
use schema::Schema;
|
||||
use std::{any::Any, collections::HashMap, sync::Arc};
|
||||
use trace::ctx::SpanContext;
|
||||
|
|
@ -192,8 +192,8 @@ mod tests {
|
|||
use arrow::record_batch::RecordBatch;
|
||||
use arrow_util::{assert_batches_eq, assert_batches_sorted_eq};
|
||||
use data_types::ColumnType;
|
||||
use iox_query::frontend::sql::SqlQueryPlanner;
|
||||
use iox_tests::util::TestCatalog;
|
||||
use query::frontend::sql::SqlQueryPlanner;
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_query() {
|
||||
|
|
|
|||
|
|
@ -1,9 +1,9 @@
|
|||
//! Ring buffer of queries that have been run with some brief information
|
||||
|
||||
use data_types::NamespaceId;
|
||||
use iox_query::QueryText;
|
||||
use iox_time::{Time, TimeProvider};
|
||||
use parking_lot::Mutex;
|
||||
use query::QueryText;
|
||||
use std::{
|
||||
collections::VecDeque,
|
||||
sync::{atomic, Arc},
|
||||
|
|
|
|||
|
|
@ -8,9 +8,9 @@ use crate::{
|
|||
};
|
||||
use backoff::{Backoff, BackoffConfig};
|
||||
use data_types::TableId;
|
||||
use iox_query::{provider::ChunkPruner, QueryChunk};
|
||||
use observability_deps::tracing::debug;
|
||||
use predicate::Predicate;
|
||||
use query::{provider::ChunkPruner, QueryChunk};
|
||||
use schema::Schema;
|
||||
use snafu::{ResultExt, Snafu};
|
||||
use std::{collections::HashMap, sync::Arc};
|
||||
|
|
|
|||
|
|
@ -9,12 +9,12 @@ use datafusion::{
|
|||
logical_plan::Expr,
|
||||
physical_plan::ExecutionPlan,
|
||||
};
|
||||
use predicate::{Predicate, PredicateBuilder};
|
||||
use query::{
|
||||
use iox_query::{
|
||||
provider::{ChunkPruner, ProviderBuilder},
|
||||
pruning::{prune_chunks, PruningObserver},
|
||||
QueryChunk,
|
||||
};
|
||||
use predicate::{Predicate, PredicateBuilder};
|
||||
use schema::Schema;
|
||||
|
||||
use super::QuerierTable;
|
||||
|
|
|
|||
|
|
@ -25,7 +25,7 @@ once_cell = { version = "1.10.0", features = ["parking_lot"] }
|
|||
predicate = { path = "../predicate" }
|
||||
schema = { path = "../schema" }
|
||||
querier = { path = "../querier" }
|
||||
query = { path = "../query" }
|
||||
iox_query = { path = "../iox_query" }
|
||||
pretty_assertions = "1.2.1"
|
||||
workspace-hack = { path = "../workspace-hack"}
|
||||
|
||||
|
|
|
|||
|
|
@ -1,8 +1,8 @@
|
|||
use std::{any::Any, sync::Arc};
|
||||
|
||||
use datafusion::catalog::catalog::CatalogProvider;
|
||||
use iox_query::{exec::ExecutionContextProvider, QueryDatabase};
|
||||
use querier::QuerierNamespace;
|
||||
use query::{exec::ExecutionContextProvider, QueryDatabase};
|
||||
|
||||
/// Abstract database used during testing.
|
||||
pub trait AbstractDb: CatalogProvider + ExecutionContextProvider + QueryDatabase {
|
||||
|
|
|
|||
|
|
@ -2,11 +2,11 @@ use crate::scenarios::*;
|
|||
use arrow::datatypes::DataType;
|
||||
use data_types::{MAX_NANO_TIME, MIN_NANO_TIME};
|
||||
use datafusion::logical_plan::{col, lit};
|
||||
use predicate::{rpc_predicate::InfluxRpcPredicate, PredicateBuilder};
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::fieldlist::{Field, FieldList},
|
||||
frontend::influxrpc::InfluxRpcPlanner,
|
||||
};
|
||||
use predicate::{rpc_predicate::InfluxRpcPredicate, PredicateBuilder};
|
||||
|
||||
/// Creates and loads several database scenarios using the db_setup
|
||||
/// function.
|
||||
|
|
|
|||
|
|
@ -16,9 +16,9 @@ use crate::{
|
|||
},
|
||||
};
|
||||
use datafusion::logical_plan::{col, lit, when};
|
||||
use iox_query::frontend::influxrpc::InfluxRpcPlanner;
|
||||
use predicate::rpc_predicate::InfluxRpcPredicate;
|
||||
use predicate::PredicateBuilder;
|
||||
use query::frontend::influxrpc::InfluxRpcPlanner;
|
||||
use test_helpers::assert_contains;
|
||||
|
||||
/// runs read_filter(predicate) and compares it to the expected
|
||||
|
|
|
|||
|
|
@ -14,9 +14,9 @@ use datafusion::{
|
|||
logical_plan::{binary_expr, Operator},
|
||||
prelude::*,
|
||||
};
|
||||
use iox_query::{frontend::influxrpc::InfluxRpcPlanner, Aggregate};
|
||||
use predicate::rpc_predicate::InfluxRpcPredicate;
|
||||
use predicate::PredicateBuilder;
|
||||
use query::{frontend::influxrpc::InfluxRpcPlanner, Aggregate};
|
||||
|
||||
/// runs read_group(predicate) and compares it to the expected
|
||||
/// output
|
||||
|
|
|
|||
|
|
@ -1,9 +1,9 @@
|
|||
//! Tests for the Influx gRPC queries
|
||||
use crate::{influxrpc::util::run_series_set_plan, scenarios::*};
|
||||
use datafusion::prelude::*;
|
||||
use iox_query::{frontend::influxrpc::InfluxRpcPlanner, Aggregate, WindowDuration};
|
||||
use predicate::rpc_predicate::InfluxRpcPredicate;
|
||||
use predicate::PredicateBuilder;
|
||||
use query::{frontend::influxrpc::InfluxRpcPlanner, Aggregate, WindowDuration};
|
||||
|
||||
/// runs read_window_aggregate(predicate) and compares it to the expected
|
||||
/// output
|
||||
|
|
|
|||
|
|
@ -2,11 +2,11 @@
|
|||
use crate::scenarios::*;
|
||||
use data_types::{MAX_NANO_TIME, MIN_NANO_TIME};
|
||||
use datafusion::logical_plan::{col, lit};
|
||||
use predicate::{rpc_predicate::InfluxRpcPredicate, PredicateBuilder};
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::stringset::{IntoStringSet, StringSetRef},
|
||||
frontend::influxrpc::InfluxRpcPlanner,
|
||||
};
|
||||
use predicate::{rpc_predicate::InfluxRpcPredicate, PredicateBuilder};
|
||||
|
||||
/// runs table_names(predicate) and compares it to the expected
|
||||
/// output
|
||||
|
|
|
|||
|
|
@ -1,11 +1,11 @@
|
|||
use crate::scenarios::*;
|
||||
use data_types::{MAX_NANO_TIME, MIN_NANO_TIME};
|
||||
use datafusion::logical_plan::{col, lit};
|
||||
use predicate::{rpc_predicate::InfluxRpcPredicate, PredicateBuilder};
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::stringset::{IntoStringSet, StringSetRef},
|
||||
frontend::influxrpc::InfluxRpcPlanner,
|
||||
};
|
||||
use predicate::{rpc_predicate::InfluxRpcPredicate, PredicateBuilder};
|
||||
|
||||
/// Creates and loads several database scenarios using the db_setup function.
|
||||
///
|
||||
|
|
|
|||
|
|
@ -1,10 +1,10 @@
|
|||
use datafusion::logical_plan::{col, lit};
|
||||
use predicate::rpc_predicate::InfluxRpcPredicate;
|
||||
use predicate::PredicateBuilder;
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::stringset::{IntoStringSet, StringSetRef},
|
||||
frontend::influxrpc::InfluxRpcPlanner,
|
||||
};
|
||||
use predicate::rpc_predicate::InfluxRpcPredicate;
|
||||
use predicate::PredicateBuilder;
|
||||
|
||||
use crate::scenarios::*;
|
||||
|
||||
|
|
|
|||
|
|
@ -1,6 +1,6 @@
|
|||
use datafusion::error::DataFusionError;
|
||||
use query::exec::IOxSessionContext;
|
||||
use query::plan::seriesset::SeriesSetPlans;
|
||||
use iox_query::exec::IOxSessionContext;
|
||||
use iox_query::plan::seriesset::SeriesSetPlans;
|
||||
|
||||
/// Run a series set plan to completion and produce a Vec<String> representation
|
||||
///
|
||||
|
|
|
|||
|
|
@ -5,7 +5,7 @@ mod setup;
|
|||
|
||||
use arrow::record_batch::RecordBatch;
|
||||
use arrow_util::{display::pretty_format_batches, test_util::sort_record_batch};
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::{Executor, ExecutorType},
|
||||
frontend::sql::SqlQueryPlanner,
|
||||
};
|
||||
|
|
@ -22,7 +22,7 @@ use self::{
|
|||
setup::TestSetup,
|
||||
};
|
||||
use crate::scenarios::{DbScenario, DbSetup};
|
||||
use query::exec::ExecutorConfig;
|
||||
use iox_query::exec::ExecutorConfig;
|
||||
|
||||
#[allow(clippy::enum_variant_names)]
|
||||
#[derive(Debug, Snafu)]
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ use super::{
|
|||
use crate::scenarios::util::{make_n_chunks_scenario, ChunkData};
|
||||
use async_trait::async_trait;
|
||||
use data_types::{DeleteExpr, DeletePredicate, Op, Scalar, TimestampRange};
|
||||
use query::frontend::sql::SqlQueryPlanner;
|
||||
use iox_query::frontend::sql::SqlQueryPlanner;
|
||||
|
||||
#[derive(Debug)]
|
||||
pub struct MeasurementWithMaxTime {}
|
||||
|
|
|
|||
|
|
@ -9,7 +9,7 @@ use super::scenarios::*;
|
|||
use arrow::record_batch::RecordBatch;
|
||||
use arrow_util::assert_batches_sorted_eq;
|
||||
use datafusion::error::DataFusionError;
|
||||
use query::frontend::sql::SqlQueryPlanner;
|
||||
use iox_query::frontend::sql::SqlQueryPlanner;
|
||||
use test_helpers::assert_contains;
|
||||
|
||||
/// Runs the query in `sql` and compares it to the expected output.
|
||||
|
|
|
|||
|
|
@ -1,7 +1,7 @@
|
|||
//! Tests for the table_names implementation
|
||||
|
||||
use arrow::datatypes::DataType;
|
||||
use query::QueryChunk;
|
||||
use iox_query::QueryChunk;
|
||||
use schema::selection::Selection;
|
||||
use schema::{builder::SchemaBuilder, sort::SortKey, Schema, TIME_COLUMN_NAME};
|
||||
|
||||
|
|
|
|||
|
|
@ -9,7 +9,7 @@ edition = "2021"
|
|||
# Workspace dependencies, in alphabetical order
|
||||
datafusion = { path = "../datafusion" }
|
||||
predicate = { path = "../predicate" }
|
||||
query = { path = "../query" }
|
||||
iox_query = { path = "../iox_query" }
|
||||
workspace-hack = { path = "../workspace-hack"}
|
||||
|
||||
# Crates.io dependencies, in alphabetical order
|
||||
|
|
|
|||
|
|
@ -5,7 +5,7 @@ pub mod planner;
|
|||
use std::sync::Arc;
|
||||
|
||||
use async_trait::async_trait;
|
||||
use query::{exec::ExecutionContextProvider, QueryDatabase};
|
||||
use iox_query::{exec::ExecutionContextProvider, QueryDatabase};
|
||||
|
||||
/// Trait that allows the query engine (which includes flight and storage/InfluxRPC) to access a virtual set of
|
||||
/// databases.
|
||||
|
|
|
|||
|
|
@ -2,7 +2,7 @@
|
|||
use std::sync::Arc;
|
||||
|
||||
use datafusion::physical_plan::ExecutionPlan;
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::IOxSessionContext,
|
||||
frontend::{influxrpc::InfluxRpcPlanner, sql::SqlQueryPlanner},
|
||||
plan::{fieldlist::FieldListPlan, seriesset::SeriesSetPlans, stringset::StringSetPlan},
|
||||
|
|
|
|||
|
|
@ -11,7 +11,7 @@ data_types = { path = "../data_types" }
|
|||
datafusion = { path = "../datafusion" }
|
||||
generated_types = { path = "../generated_types" }
|
||||
observability_deps = { path = "../observability_deps" }
|
||||
query = { path = "../query" }
|
||||
iox_query = { path = "../iox_query" }
|
||||
service_common = { path = "../service_common" }
|
||||
|
||||
# Crates.io dependencies, in alphabetical order
|
||||
|
|
|
|||
|
|
@ -16,13 +16,13 @@ use data_types::{DatabaseName, DatabaseNameError};
|
|||
use datafusion::physical_plan::ExecutionPlan;
|
||||
use futures::{SinkExt, Stream, StreamExt};
|
||||
use generated_types::influxdata::iox::querier::v1 as proto;
|
||||
use observability_deps::tracing::{info, warn};
|
||||
use pin_project::{pin_project, pinned_drop};
|
||||
use prost::Message;
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::{ExecutionContextProvider, IOxSessionContext},
|
||||
QueryCompletedToken, QueryDatabase,
|
||||
};
|
||||
use observability_deps::tracing::{info, warn};
|
||||
use pin_project::{pin_project, pinned_drop};
|
||||
use prost::Message;
|
||||
use serde::Deserialize;
|
||||
use service_common::{planner::Planner, QueryDatabaseProvider};
|
||||
use snafu::{ResultExt, Snafu};
|
||||
|
|
|
|||
|
|
@ -10,7 +10,7 @@ datafusion = { path = "../datafusion" }
|
|||
generated_types = { path = "../generated_types" }
|
||||
observability_deps = { path = "../observability_deps" }
|
||||
predicate = { path = "../predicate" }
|
||||
query = { path = "../query" }
|
||||
iox_query = { path = "../iox_query" }
|
||||
query_functions = { path = "../query_functions"}
|
||||
schema = { path = "../schema" }
|
||||
service_common = { path = "../service_common" }
|
||||
|
|
|
|||
|
|
@ -5,12 +5,12 @@ use std::{collections::BTreeSet, fmt, sync::Arc};
|
|||
|
||||
use arrow::datatypes::DataType as ArrowDataType;
|
||||
|
||||
use observability_deps::tracing::trace;
|
||||
use predicate::rpc_predicate::{FIELD_COLUMN_NAME, MEASUREMENT_COLUMN_NAME};
|
||||
use query::exec::{
|
||||
use iox_query::exec::{
|
||||
fieldlist::FieldList,
|
||||
seriesset::series::{self, Either},
|
||||
};
|
||||
use observability_deps::tracing::trace;
|
||||
use predicate::rpc_predicate::{FIELD_COLUMN_NAME, MEASUREMENT_COLUMN_NAME};
|
||||
|
||||
use generated_types::{
|
||||
measurement_fields_response::{FieldType, MessageField},
|
||||
|
|
@ -28,7 +28,7 @@ use snafu::Snafu;
|
|||
pub enum Error {
|
||||
#[snafu(display("Error converting series set to gRPC: {}", source))]
|
||||
ConvertingSeries {
|
||||
source: query::exec::seriesset::series::Error,
|
||||
source: iox_query::exec::seriesset::series::Error,
|
||||
},
|
||||
|
||||
#[snafu(display("Unsupported field data type in gRPC data translation: {}", data_type))]
|
||||
|
|
@ -324,7 +324,7 @@ mod tests {
|
|||
datatypes::DataType as ArrowDataType,
|
||||
record_batch::RecordBatch,
|
||||
};
|
||||
use query::exec::{
|
||||
use iox_query::exec::{
|
||||
field::FieldIndexes,
|
||||
fieldlist::Field,
|
||||
seriesset::{
|
||||
|
|
|
|||
|
|
@ -23,12 +23,12 @@ use generated_types::{
|
|||
};
|
||||
|
||||
use super::{TAG_KEY_FIELD, TAG_KEY_MEASUREMENT};
|
||||
use iox_query::{Aggregate as QueryAggregate, WindowDuration};
|
||||
use observability_deps::tracing::warn;
|
||||
use predicate::{
|
||||
rpc_predicate::{InfluxRpcPredicate, FIELD_COLUMN_NAME, MEASUREMENT_COLUMN_NAME},
|
||||
PredicateBuilder,
|
||||
};
|
||||
use query::{Aggregate as QueryAggregate, WindowDuration};
|
||||
use snafu::{OptionExt, ResultExt, Snafu};
|
||||
|
||||
#[derive(Debug, Snafu)]
|
||||
|
|
|
|||
|
|
@ -23,14 +23,14 @@ use generated_types::{
|
|||
TagValuesGroupedByMeasurementAndTagKeyRequest, TagValuesRequest, TagValuesResponse,
|
||||
TimestampRange,
|
||||
};
|
||||
use observability_deps::tracing::{error, info, trace};
|
||||
use query::{
|
||||
use iox_query::{
|
||||
exec::{
|
||||
fieldlist::FieldList, seriesset::converter::Error as SeriesSetError,
|
||||
ExecutionContextProvider, IOxSessionContext,
|
||||
},
|
||||
QueryDatabase, QueryText,
|
||||
};
|
||||
use observability_deps::tracing::{error, info, trace};
|
||||
use service_common::{planner::Planner, QueryDatabaseProvider};
|
||||
use snafu::{OptionExt, ResultExt, Snafu};
|
||||
use std::{
|
||||
|
|
@ -1398,14 +1398,14 @@ mod tests {
|
|||
generated_types::*,
|
||||
Client as StorageClient, OrgAndBucket,
|
||||
};
|
||||
use iox_query::{
|
||||
exec::Executor,
|
||||
test::{TestChunk, TestDatabase},
|
||||
};
|
||||
use metric::{Attributes, Metric, U64Counter};
|
||||
use panic_logging::SendPanicsToTracing;
|
||||
use parking_lot::Mutex;
|
||||
use predicate::{PredicateBuilder, PredicateMatch};
|
||||
use query::{
|
||||
exec::Executor,
|
||||
test::{TestChunk, TestDatabase},
|
||||
};
|
||||
use service_common::QueryDatabaseProvider;
|
||||
use std::{
|
||||
collections::BTreeMap,
|
||||
|
|
|
|||
Loading…
Reference in New Issue