ci: fix cargo deny (#4629)

* ci: fix cargo deny

* chore: downgrade `socket2`, version 0.4.5 was yanked

* chore: rename `query` to `iox_query`

`query` is already taken on crates.io and yanked and I am getting tired
of working around that.
pull/24376/head
Marco Neumann 2022-05-18 11:38:35 +02:00 committed by GitHub
parent 7f60f69734
commit 52346642a0
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
98 changed files with 168 additions and 166 deletions

View File

@ -131,7 +131,7 @@ jobs:
# https://github.com/EmbarkStudios/cargo-deny/issues/388
name: Pull advisory-db
command: |
git clone --depth=1 https://github.com/RustSec/advisory-db.git /usr/local/cargo/advisory-dbs/github.com-dfdf83aac62a7e5e
git clone --depth=1 https://github.com/RustSec/advisory-db.git /usr/local/cargo/advisory-dbs/github.com-2f857891b7f43c59
- run:
name: cargo-deny Checks
command: cargo deny check -s --disable-fetch

88
Cargo.lock generated
View File

@ -845,6 +845,7 @@ dependencies = [
"datafusion 0.1.0",
"futures",
"iox_catalog",
"iox_query",
"iox_tests",
"iox_time",
"metric",
@ -853,7 +854,6 @@ dependencies = [
"parquet_file",
"predicate",
"querier",
"query",
"schema",
"snafu",
"test_helpers",
@ -2181,6 +2181,7 @@ dependencies = [
"influxdb_storage_client",
"influxrpc_parser",
"iox_catalog",
"iox_query",
"iox_time",
"ioxd_common",
"ioxd_compactor",
@ -2200,7 +2201,6 @@ dependencies = [
"parquet_file",
"predicate",
"predicates",
"query",
"rustyline",
"schema",
"serde_json",
@ -2312,6 +2312,7 @@ dependencies = [
"generated_types",
"hyper",
"iox_catalog",
"iox_query",
"iox_time",
"lazy_static",
"metric",
@ -2325,7 +2326,6 @@ dependencies = [
"pin-project",
"predicate",
"prost",
"query",
"schema",
"snafu",
"test_helpers",
@ -2447,6 +2447,35 @@ dependencies = [
"workspace-hack",
]
[[package]]
name = "iox_query"
version = "0.1.0"
dependencies = [
"arrow",
"arrow_util",
"async-trait",
"chrono",
"croaring",
"data_types",
"datafusion 0.1.0",
"datafusion_util",
"executor",
"futures",
"hashbrown 0.12.1",
"itertools",
"observability_deps",
"parking_lot 0.12.0",
"predicate",
"query_functions",
"schema",
"snafu",
"test_helpers",
"tokio",
"tokio-stream",
"trace",
"workspace-hack",
]
[[package]]
name = "iox_tests"
version = "0.1.0"
@ -2456,12 +2485,12 @@ dependencies = [
"data_types",
"datafusion 0.1.0",
"iox_catalog",
"iox_query",
"iox_time",
"metric",
"mutable_batch_lp",
"object_store",
"parquet_file",
"query",
"schema",
"uuid 0.8.2",
"workspace-hack",
@ -2530,11 +2559,11 @@ dependencies = [
"data_types",
"hyper",
"iox_catalog",
"iox_query",
"iox_time",
"ioxd_common",
"metric",
"object_store",
"query",
"thiserror",
"trace",
"workspace-hack",
@ -2550,10 +2579,10 @@ dependencies = [
"hyper",
"ingester",
"iox_catalog",
"iox_query",
"ioxd_common",
"metric",
"object_store",
"query",
"thiserror",
"trace",
"workspace-hack",
@ -2570,13 +2599,13 @@ dependencies = [
"generated_types",
"hyper",
"iox_catalog",
"iox_query",
"iox_tests",
"iox_time",
"ioxd_common",
"metric",
"object_store",
"querier",
"query",
"service_grpc_flight",
"service_grpc_influxrpc",
"tokio",
@ -4132,6 +4161,7 @@ dependencies = [
"generated_types",
"influxdb_iox_client",
"iox_catalog",
"iox_query",
"iox_tests",
"iox_time",
"metric",
@ -4142,7 +4172,6 @@ dependencies = [
"parquet_file",
"pin-project",
"predicate",
"query",
"rand",
"schema",
"service_common",
@ -4158,35 +4187,6 @@ dependencies = [
"workspace-hack",
]
[[package]]
name = "query"
version = "0.1.0"
dependencies = [
"arrow",
"arrow_util",
"async-trait",
"chrono",
"croaring",
"data_types",
"datafusion 0.1.0",
"datafusion_util",
"executor",
"futures",
"hashbrown 0.12.1",
"itertools",
"observability_deps",
"parking_lot 0.12.0",
"predicate",
"query_functions",
"schema",
"snafu",
"test_helpers",
"tokio",
"tokio-stream",
"trace",
"workspace-hack",
]
[[package]]
name = "query_functions"
version = "0.1.0"
@ -4222,6 +4222,7 @@ dependencies = [
"influxdb_iox_client",
"ingester",
"iox_catalog",
"iox_query",
"iox_tests",
"itertools",
"mutable_batch",
@ -4230,7 +4231,6 @@ dependencies = [
"predicate",
"pretty_assertions",
"querier",
"query",
"schema",
"snafu",
"tempfile",
@ -5033,8 +5033,8 @@ version = "0.1.0"
dependencies = [
"async-trait",
"datafusion 0.1.0",
"iox_query",
"predicate",
"query",
"workspace-hack",
]
@ -5064,10 +5064,10 @@ dependencies = [
"datafusion 0.1.0",
"futures",
"generated_types",
"iox_query",
"observability_deps",
"pin-project",
"prost",
"query",
"serde",
"serde_json",
"service_common",
@ -5088,13 +5088,13 @@ dependencies = [
"futures",
"generated_types",
"influxdb_storage_client",
"iox_query",
"metric",
"observability_deps",
"panic_logging",
"parking_lot 0.12.0",
"predicate",
"prost",
"query",
"query_functions",
"regex",
"schema",
@ -5291,12 +5291,12 @@ checksum = "45456094d1983e2ee2a18fdfebce3189fa451699d0502cb8e3b49dba5ba41451"
[[package]]
name = "socket2"
version = "0.4.5"
version = "0.4.4"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "ca642ba17f8b2995138b1d7711829c92e98c0a25ea019de790f4f09279c4e296"
checksum = "66d72b759436ae32898a2af0a14218dbf55efde3feeb170eb623637db85ee1e0"
dependencies = [
"libc",
"windows-sys 0.36.1",
"winapi",
]
[[package]]

View File

@ -26,6 +26,7 @@ members = [
"iox_catalog",
"iox_data_generator",
"iox_gitops_adapter",
"iox_query",
"iox_tests",
"iox_time",
"ioxd_common",
@ -48,7 +49,6 @@ members = [
"parquet_file",
"predicate",
"querier",
"query",
"query_functions",
"query_tests",
"read_buffer",

View File

@ -18,7 +18,7 @@ object_store = "0.0.1"
observability_deps = { path = "../observability_deps" }
parquet_file = { path = "../parquet_file" }
predicate = { path = "../predicate" }
query = { path = "../query" }
iox_query = { path = "../iox_query" }
schema = { path = "../schema" }
snafu = "0.7"
thiserror = "1.0"

View File

@ -17,6 +17,13 @@ use data_types::{
};
use datafusion::error::DataFusionError;
use iox_catalog::interface::{Catalog, Transaction, INITIAL_COMPACTION_LEVEL};
use iox_query::{
exec::{Executor, ExecutorType},
frontend::reorg::ReorgPlanner,
provider::overlap::group_potential_duplicates,
util::compute_timenanosecond_min_max,
QueryChunk,
};
use iox_time::{Time, TimeProvider};
use metric::{Attributes, Metric, U64Counter, U64Gauge, U64Histogram, U64HistogramOptions};
use object_store::DynObjectStore;
@ -25,13 +32,6 @@ use parquet_file::{
metadata::{IoxMetadata, IoxParquetMetaData},
ParquetFilePath,
};
use query::{
exec::{Executor, ExecutorType},
frontend::reorg::ReorgPlanner,
provider::overlap::group_potential_duplicates,
util::compute_timenanosecond_min_max,
QueryChunk,
};
use schema::sort::SortKey;
use snafu::{ensure, OptionExt, ResultExt, Snafu};
use std::{
@ -87,7 +87,7 @@ pub enum Error {
#[snafu(display("Error building compact logical plan {}", source))]
CompactLogicalPlan {
source: query::frontend::reorg::Error,
source: iox_query::frontend::reorg::Error,
},
#[snafu(display("Error building compact physical plan {}", source))]
@ -103,7 +103,7 @@ pub enum Error {
RowCountTypeConversion { source: std::num::TryFromIntError },
#[snafu(display("Error computing min and max for record batches: {}", source))]
MinMax { source: query::util::Error },
MinMax { source: iox_query::util::Error },
#[snafu(display("Error while starting catalog transaction {}", source))]
Transaction {
@ -1108,13 +1108,13 @@ mod tests {
use data_types::{ChunkId, KafkaPartition, NamespaceId, ParquetFileParams, SequenceNumber};
use futures::TryStreamExt;
use iox_catalog::interface::INITIAL_COMPACTION_LEVEL;
use iox_query::test::{raw_data, TestChunk};
use iox_tests::util::TestCatalog;
use iox_time::SystemProvider;
use querier::{
cache::CatalogCache,
chunk::{collect_read_filter, ParquetChunkAdapter},
};
use query::test::{raw_data, TestChunk};
use std::sync::atomic::{AtomicI64, Ordering};
// Simulate unique ID generation

View File

@ -8,10 +8,10 @@ use futures::{
select, FutureExt, TryFutureExt,
};
use iox_catalog::interface::Catalog;
use iox_query::exec::Executor;
use iox_time::TimeProvider;
use object_store::DynObjectStore;
use observability_deps::tracing::*;
use query::exec::Executor;
use std::sync::Arc;
use thiserror::Error;
use tokio::task::{JoinError, JoinHandle};

View File

@ -5,13 +5,13 @@ use data_types::{
TimestampMinMax, Tombstone,
};
use datafusion::physical_plan::SendableRecordBatchStream;
use observability_deps::tracing::trace;
use parquet_file::chunk::ParquetChunk;
use predicate::{delete_predicate::tombstones_to_delete_predicates, Predicate, PredicateMatch};
use query::{
use iox_query::{
exec::{stringset::StringSet, IOxSessionContext},
QueryChunk, QueryChunkError, QueryChunkMeta,
};
use observability_deps::tracing::trace;
use parquet_file::chunk::ParquetChunk;
use predicate::{delete_predicate::tombstones_to_delete_predicates, Predicate, PredicateMatch};
use schema::{merge::SchemaMerger, selection::Selection, sort::SortKey, Schema};
use snafu::{ResultExt, Snafu};
use std::sync::Arc;

View File

@ -26,6 +26,8 @@ ignore = [
# why needed: used by `criterion`
# upstream issue: https://github.com/bheisler/criterion.rs/issues/534
"RUSTSEC-2021-0127",
"query",
]
[licenses]

View File

@ -43,5 +43,5 @@ We hold monthly Tech Talks that explain the project's technical underpinnings. Y
* [Data Deduplication & Sort Key](dedup_and_sort.md)
* Transactions and locks (to be written & linked)
* [Thoughts on using multiple cores / thread pools](multi_core_tasks.md)
* [Query Engine Docs](../query/README.md)
* [Query Engine Docs](../iox_query/README.md)
* [Notes on the use of local filesystems](local_filesystems.md)

View File

@ -27,7 +27,7 @@ object_store_metrics = { path = "../object_store_metrics" }
observability_deps = { path = "../observability_deps" }
panic_logging = { path = "../panic_logging" }
parquet_file = { path = "../parquet_file" }
query = { path = "../query" }
iox_query = { path = "../iox_query" }
schema = { path = "../schema" }
iox_time = { path = "../iox_time" }
trace_exporters = { path = "../trace_exporters" }

View File

@ -8,6 +8,7 @@ use clap_blocks::{
object_store::ObjectStoreConfig, run_config::RunConfig, socket_addr::SocketAddr,
write_buffer::WriteBufferConfig,
};
use iox_query::exec::Executor;
use iox_time::{SystemProvider, TimeProvider};
use ioxd_common::{
server_type::{CommonServerState, CommonServerStateError},
@ -19,7 +20,6 @@ use ioxd_querier::{create_querier_server_type, QuerierServerTypeArgs};
use ioxd_router::create_router_server_type;
use object_store::DynObjectStore;
use observability_deps::tracing::*;
use query::exec::Executor;
use std::{path::PathBuf, sync::Arc};
use thiserror::Error;
use trace_exporters::TracingConfig;

View File

@ -1,10 +1,10 @@
//! Implementation of command line option for running the compactor
use iox_query::exec::Executor;
use iox_time::{SystemProvider, TimeProvider};
use object_store::DynObjectStore;
use object_store_metrics::ObjectStoreMetrics;
use observability_deps::tracing::*;
use query::exec::Executor;
use std::sync::Arc;
use thiserror::Error;

View File

@ -5,6 +5,7 @@ use clap_blocks::{
catalog_dsn::CatalogDsnConfig, ingester::IngesterConfig, run_config::RunConfig,
write_buffer::WriteBufferConfig,
};
use iox_query::exec::Executor;
use iox_time::{SystemProvider, TimeProvider};
use ioxd_common::server_type::{CommonServerState, CommonServerStateError};
use ioxd_common::Service;
@ -12,7 +13,6 @@ use ioxd_ingester::create_ingester_server_type;
use object_store::DynObjectStore;
use object_store_metrics::ObjectStoreMetrics;
use observability_deps::tracing::*;
use query::exec::Executor;
use std::sync::Arc;
use thiserror::Error;

View File

@ -1,11 +1,11 @@
//! Implementation of command line option for running the querier
use clap_blocks::querier::QuerierConfig;
use iox_query::exec::Executor;
use iox_time::{SystemProvider, TimeProvider};
use object_store::DynObjectStore;
use object_store_metrics::ObjectStoreMetrics;
use observability_deps::tracing::*;
use query::exec::Executor;
use std::sync::Arc;
use thiserror::Error;

View File

@ -30,7 +30,7 @@ parquet_file = { path = "../parquet_file" }
pin-project = "1.0"
predicate = { path = "../predicate" }
prost = "0.10"
query = { path = "../query" }
iox_query = { path = "../iox_query" }
schema = { path = "../schema" }
snafu = "0.7"
thiserror = "1.0"

View File

@ -5,14 +5,14 @@ use arrow::record_batch::RecordBatch;
use data_types::{NamespaceId, PartitionInfo};
use datafusion::{error::DataFusionError, physical_plan::SendableRecordBatchStream};
use iox_catalog::interface::INITIAL_COMPACTION_LEVEL;
use iox_time::{Time, TimeProvider};
use parquet_file::metadata::IoxMetadata;
use query::{
use iox_query::{
exec::{Executor, ExecutorType},
frontend::reorg::ReorgPlanner,
util::compute_timenanosecond_min_max,
QueryChunk, QueryChunkMeta,
};
use iox_time::{Time, TimeProvider};
use parquet_file::metadata::IoxMetadata;
use schema::sort::{adjust_sort_key_columns, compute_sort_key, SortKey};
use snafu::{ResultExt, Snafu};
use std::sync::Arc;
@ -22,7 +22,7 @@ use std::sync::Arc;
pub enum Error {
#[snafu(display("Error while building logical plan for Ingester's compaction"))]
LogicalPlan {
source: query::frontend::reorg::Error,
source: iox_query::frontend::reorg::Error,
},
#[snafu(display("Error while building physical plan for Ingester's compaction"))]
@ -46,7 +46,7 @@ pub enum Error {
RowCountTypeConversion { source: std::num::TryFromIntError },
#[snafu(display("Error computing min and max for record batches: {}", source))]
MinMax { source: query::util::Error },
MinMax { source: iox_query::util::Error },
}
/// A specialized `Error` for Ingester's Compact errors

View File

@ -17,6 +17,7 @@ use data_types::{
use datafusion::physical_plan::SendableRecordBatchStream;
use dml::DmlOperation;
use iox_catalog::interface::Catalog;
use iox_query::exec::Executor;
use iox_time::SystemProvider;
use metric::U64Counter;
use mutable_batch::MutableBatch;
@ -24,7 +25,6 @@ use object_store::DynObjectStore;
use observability_deps::tracing::{debug, warn};
use parking_lot::RwLock;
use predicate::Predicate;
use query::exec::Executor;
use schema::{selection::Selection, Schema};
use snafu::{OptionExt, ResultExt, Snafu};
use std::{

View File

@ -21,11 +21,11 @@ use futures::{
};
use generated_types::ingester::IngesterQueryRequest;
use iox_catalog::interface::Catalog;
use iox_query::exec::Executor;
use iox_time::{SystemProvider, TimeProvider};
use metric::{Metric, U64Histogram, U64HistogramOptions};
use object_store::DynObjectStore;
use observability_deps::tracing::*;
use query::exec::Executor;
use snafu::{ResultExt, Snafu};
use std::{collections::BTreeMap, sync::Arc, time::Duration};
use tokio::task::{JoinError, JoinHandle};

View File

@ -84,9 +84,9 @@ mod tests {
use data_types::{NamespaceId, PartitionId, SequenceNumber, SequencerId, TableId};
use futures::{StreamExt, TryStreamExt};
use iox_catalog::interface::INITIAL_COMPACTION_LEVEL;
use iox_query::test::{raw_data, TestChunk};
use iox_time::Time;
use object_store::memory::InMemory;
use query::test::{raw_data, TestChunk};
use std::sync::Arc;
use uuid::Uuid;

View File

@ -14,13 +14,13 @@ use datafusion::{
},
};
use generated_types::ingester::IngesterQueryRequest;
use observability_deps::tracing::debug;
use predicate::Predicate;
use query::{
use iox_query::{
exec::{Executor, ExecutorType},
frontend::reorg::ReorgPlanner,
QueryChunkMeta,
};
use observability_deps::tracing::debug;
use predicate::Predicate;
use schema::{
merge::{merge_record_batch_schemas, SchemaMerger},
selection::Selection,
@ -37,7 +37,7 @@ pub enum Error {
#[snafu(display("Error building logical plan for querying Ingester data to send to Querier"))]
LogicalPlan {
source: query::frontend::reorg::Error,
source: iox_query::frontend::reorg::Error,
},
#[snafu(display(

View File

@ -16,16 +16,16 @@ use datafusion::{
},
};
use datafusion_util::batch_filter;
use iox_query::{
exec::{stringset::StringSet, IOxSessionContext},
util::{df_physical_expr_from_schema_and_expr, MissingColumnsToNull},
QueryChunk, QueryChunkError, QueryChunkMeta,
};
use observability_deps::tracing::{debug, trace};
use predicate::{
delete_predicate::{tombstones_to_delete_predicates, tombstones_to_delete_predicates_iter},
Predicate, PredicateMatch,
};
use query::{
exec::{stringset::StringSet, IOxSessionContext},
util::{df_physical_expr_from_schema_and_expr, MissingColumnsToNull},
QueryChunk, QueryChunkError, QueryChunkMeta,
};
use schema::{merge::merge_record_batch_schemas, selection::Selection, sort::SortKey, Schema};
use snafu::{ResultExt, Snafu};
use std::sync::Arc;

View File

@ -20,10 +20,10 @@ use iox_catalog::{
interface::{Catalog, INITIAL_COMPACTION_LEVEL},
mem::MemCatalog,
};
use iox_query::test::{raw_data, TestChunk};
use iox_time::{SystemProvider, Time, TimeProvider};
use object_store::memory::InMemory;
use parquet_file::metadata::IoxMetadata;
use query::test::{raw_data, TestChunk};
use schema::sort::SortKey;
use std::{collections::BTreeMap, sync::Arc};
use uuid::Uuid;
@ -677,7 +677,7 @@ pub fn make_ingester_data(two_partitions: bool, loc: DataLocation) -> IngesterDa
let metrics: Arc<metric::Registry> = Default::default();
let catalog: Arc<dyn Catalog> = Arc::new(MemCatalog::new(Arc::clone(&metrics)));
let object_store = Arc::new(InMemory::new());
let exec = Arc::new(query::exec::Executor::new(1));
let exec = Arc::new(iox_query::exec::Executor::new(1));
// Make data for one sequencer/shard and two tables
let seq_id = SequencerId::new(1);
@ -730,7 +730,7 @@ pub async fn make_ingester_data_with_tombstones(loc: DataLocation) -> IngesterDa
let metrics: Arc<metric::Registry> = Default::default();
let catalog: Arc<dyn Catalog> = Arc::new(MemCatalog::new(metrics));
let object_store = Arc::new(InMemory::new());
let exec = Arc::new(query::exec::Executor::new(1));
let exec = Arc::new(iox_query::exec::Executor::new(1));
// Make data for one sequencer/shard and two tables
let seq_id = SequencerId::new(1);
@ -845,7 +845,7 @@ pub(crate) fn make_partitions(
/// Make data for one partition with tombstones
pub(crate) async fn make_one_partition_with_tombstones(
exec: &query::exec::Executor,
exec: &iox_query::exec::Executor,
loc: DataLocation,
sequencer_id: SequencerId,
table_id: TableId,

View File

@ -1,5 +1,5 @@
[package]
name = "query"
name = "iox_query"
version = "0.1.0"
authors = ["Andrew Lamb <andrew@nerdnetworks.org>"]
edition = "2021"

View File

@ -383,7 +383,7 @@ pub fn compute_sort_key<'a>(summaries: impl Iterator<Item = &'a TableSummary>) -
// error[E0433]: failed to resolve: could not find `test` in `storage`
// --> src/server/mutable_buffer_routes.rs:353:19
// |
// 353 | use query::test::TestDatabaseStore;
// 353 | use iox_query::test::TestDatabaseStore;
// | ^^^^ could not find `test` in `query`
//

View File

@ -16,7 +16,7 @@ metric = { path = "../metric" }
mutable_batch_lp = { path = "../mutable_batch_lp" }
object_store = "0.0.1"
parquet_file = { path = "../parquet_file" }
query = { path = "../query" }
iox_query = { path = "../iox_query" }
schema = { path = "../schema" }
uuid = { version = "0.8", features = ["v4"] }
workspace-hack = { path = "../workspace-hack"}

View File

@ -15,6 +15,7 @@ use iox_catalog::{
interface::{Catalog, PartitionRepo, INITIAL_COMPACTION_LEVEL},
mem::MemCatalog,
};
use iox_query::{exec::Executor, provider::RecordBatchDeduplicator, util::arrow_sort_key_exprs};
use iox_time::{MockProvider, Time, TimeProvider};
use mutable_batch_lp::test_helpers::lp_to_mutable_batch;
use object_store::{memory::InMemory, DynObjectStore};
@ -22,7 +23,6 @@ use parquet_file::{
metadata::{IoxMetadata, IoxParquetMetaData},
ParquetFilePath,
};
use query::{exec::Executor, provider::RecordBatchDeduplicator, util::arrow_sort_key_exprs};
use schema::{
selection::Selection,
sort::{adjust_sort_key_columns, SortKey, SortKeyBuilder},

View File

@ -13,7 +13,7 @@ data_types = { path = "../data_types" }
iox_catalog = { path = "../iox_catalog" }
ioxd_common = { path = "../ioxd_common" }
metric = { path = "../metric" }
query = { path = "../query" }
iox_query = { path = "../iox_query" }
object_store = "0.0.1"
iox_time = { path = "../iox_time" }
trace = { path = "../trace" }

View File

@ -7,6 +7,7 @@ use compactor::{
use data_types::KafkaPartition;
use hyper::{Body, Request, Response};
use iox_catalog::interface::Catalog;
use iox_query::exec::Executor;
use iox_time::TimeProvider;
use ioxd_common::{
add_service,
@ -18,7 +19,6 @@ use ioxd_common::{
};
use metric::Registry;
use object_store::DynObjectStore;
use query::exec::Executor;
use std::{
fmt::{Debug, Display},
sync::Arc,

View File

@ -12,7 +12,7 @@ iox_catalog = { path = "../iox_catalog" }
ioxd_common = { path = "../ioxd_common" }
metric = { path = "../metric" }
object_store = "0.0.1"
query = { path = "../query" }
iox_query = { path = "../iox_query" }
trace = { path = "../trace" }
write_buffer = { path = "../write_buffer" }

View File

@ -8,6 +8,7 @@ use ingester::{
server::{grpc::GrpcDelegate, http::HttpDelegate, IngesterServer},
};
use iox_catalog::interface::Catalog;
use iox_query::exec::Executor;
use ioxd_common::{
add_service,
http::error::{HttpApiError, HttpApiErrorCode, HttpApiErrorSource},
@ -18,7 +19,6 @@ use ioxd_common::{
};
use metric::Registry;
use object_store::DynObjectStore;
use query::exec::Executor;
use std::{
collections::BTreeMap,
fmt::{Debug, Display},

View File

@ -14,7 +14,7 @@ ioxd_common = { path = "../ioxd_common" }
metric = { path = "../metric" }
object_store = "0.0.1"
querier = { path = "../querier" }
query = { path = "../query" }
iox_query = { path = "../iox_query" }
service_grpc_flight = { path = "../service_grpc_flight" }
service_grpc_influxrpc = { path = "../service_grpc_influxrpc" }
iox_time = { path = "../iox_time" }

View File

@ -6,6 +6,7 @@ use std::{
use async_trait::async_trait;
use hyper::{Body, Request, Response};
use iox_catalog::interface::Catalog;
use iox_query::exec::Executor;
use iox_time::TimeProvider;
use metric::Registry;
use object_store::DynObjectStore;
@ -13,7 +14,6 @@ use querier::{
create_ingester_connection, QuerierCatalogCache, QuerierDatabase, QuerierHandler,
QuerierHandlerImpl, QuerierServer,
};
use query::exec::Executor;
use trace::TraceCollector;
use ioxd_common::{

View File

@ -23,7 +23,7 @@ parking_lot = "0.12"
parquet_file = { path = "../parquet_file" }
pin-project = "1.0"
predicate = { path = "../predicate" }
query = { path = "../query" }
iox_query = { path = "../iox_query" }
rand = "0.8.3"
service_common = { path = "../service_common" }
service_grpc_schema = { path = "../service_grpc_schema" }

View File

@ -8,12 +8,12 @@ use data_types::{
};
use futures::StreamExt;
use iox_catalog::interface::Catalog;
use iox_query::{exec::IOxSessionContext, QueryChunk};
use iox_time::TimeProvider;
use object_store::DynObjectStore;
use parquet_file::chunk::{
new_parquet_chunk, ChunkMetrics as ParquetChunkMetrics, DecodedParquetFile, ParquetChunk,
};
use query::{exec::IOxSessionContext, QueryChunk};
use schema::{selection::Selection, sort::SortKey};
use std::sync::Arc;
use uuid::Uuid;
@ -292,8 +292,8 @@ pub mod tests {
use super::*;
use arrow::datatypes::DataType;
use arrow_util::assert_batches_eq;
use iox_query::QueryChunkMeta;
use iox_tests::util::TestCatalog;
use query::QueryChunkMeta;
use schema::{builder::SchemaBuilder, sort::SortKeyBuilder};
#[tokio::test]

View File

@ -2,9 +2,9 @@ use crate::chunk::{ChunkStorage, QuerierChunk};
use data_types::{
ChunkId, ChunkOrder, DeletePredicate, PartitionId, TableSummary, TimestampMinMax,
};
use iox_query::{QueryChunk, QueryChunkError, QueryChunkMeta};
use observability_deps::tracing::debug;
use predicate::PredicateMatch;
use query::{QueryChunk, QueryChunkError, QueryChunkMeta};
use schema::{sort::SortKey, Schema};
use snafu::{ResultExt, Snafu};
use std::sync::Arc;
@ -86,10 +86,10 @@ impl QueryChunk for QuerierChunk {
fn column_names(
&self,
_ctx: query::exec::IOxSessionContext,
_ctx: iox_query::exec::IOxSessionContext,
predicate: &predicate::Predicate,
columns: schema::selection::Selection<'_>,
) -> Result<Option<query::exec::stringset::StringSet>, QueryChunkError> {
) -> Result<Option<iox_query::exec::stringset::StringSet>, QueryChunkError> {
match &self.storage {
ChunkStorage::Parquet { chunk, .. } => {
if !predicate.is_empty() {
@ -103,10 +103,10 @@ impl QueryChunk for QuerierChunk {
fn column_values(
&self,
_ctx: query::exec::IOxSessionContext,
_ctx: iox_query::exec::IOxSessionContext,
_column_name: &str,
_predicate: &predicate::Predicate,
) -> Result<Option<query::exec::stringset::StringSet>, QueryChunkError> {
) -> Result<Option<iox_query::exec::stringset::StringSet>, QueryChunkError> {
match &self.storage {
ChunkStorage::Parquet { .. } => {
// Since DataFusion can read Parquet, there is no advantage to
@ -118,7 +118,7 @@ impl QueryChunk for QuerierChunk {
fn read_filter(
&self,
mut ctx: query::exec::IOxSessionContext,
mut ctx: iox_query::exec::IOxSessionContext,
predicate: &predicate::Predicate,
selection: schema::selection::Selection<'_>,
) -> Result<datafusion::physical_plan::SendableRecordBatchStream, QueryChunkError> {

View File

@ -7,9 +7,9 @@ use crate::{
use async_trait::async_trait;
use backoff::{Backoff, BackoffConfig};
use data_types::Namespace;
use iox_query::exec::Executor;
use object_store::DynObjectStore;
use parking_lot::RwLock;
use query::exec::Executor;
use service_common::QueryDatabaseProvider;
use std::{collections::HashMap, sync::Arc};

View File

@ -131,9 +131,9 @@ mod tests {
use std::time::Duration;
use iox_catalog::mem::MemCatalog;
use iox_query::exec::Executor;
use iox_time::{MockProvider, Time};
use object_store::memory::InMemory;
use query::exec::Executor;
use crate::{cache::CatalogCache, create_ingester_connection_for_testing};

View File

@ -16,13 +16,13 @@ use generated_types::{
influxdata::iox::ingester::v1::GetWriteInfoResponse, ingester::IngesterQueryRequest,
write_info::merge_responses,
};
use observability_deps::tracing::{debug, trace};
use predicate::{Predicate, PredicateMatch};
use query::{
use iox_query::{
exec::{stringset::StringSet, IOxSessionContext},
util::compute_timenanosecond_min_max,
QueryChunk, QueryChunkError, QueryChunkMeta,
};
use observability_deps::tracing::{debug, trace};
use predicate::{Predicate, PredicateMatch};
use schema::{selection::Selection, sort::SortKey, InfluxColumnType, InfluxFieldType, Schema};
use snafu::{ensure, OptionExt, ResultExt, Snafu};
use std::{any::Any, collections::HashMap, sync::Arc};

View File

@ -6,8 +6,8 @@ use crate::{
};
use backoff::BackoffConfig;
use data_types::{NamespaceId, NamespaceSchema};
use iox_query::exec::Executor;
use object_store::DynObjectStore;
use query::exec::Executor;
use schema::Schema;
use std::{collections::HashMap, sync::Arc};

View File

@ -1,4 +1,4 @@
//! This module contains implementations of [query] interfaces for [QuerierNamespace].
//! This module contains implementations of [`iox_query`] interfaces for [QuerierNamespace].
use crate::{
namespace::QuerierNamespace,
@ -12,12 +12,12 @@ use datafusion::{
catalog::{catalog::CatalogProvider, schema::SchemaProvider},
datasource::TableProvider,
};
use observability_deps::tracing::{debug, trace};
use predicate::{rpc_predicate::QueryDatabaseMeta, Predicate};
use query::{
use iox_query::{
exec::{ExecutionContextProvider, ExecutorType, IOxSessionContext},
QueryChunk, QueryCompletedToken, QueryDatabase, QueryDatabaseError, QueryText, DEFAULT_SCHEMA,
};
use observability_deps::tracing::{debug, trace};
use predicate::{rpc_predicate::QueryDatabaseMeta, Predicate};
use schema::Schema;
use std::{any::Any, collections::HashMap, sync::Arc};
use trace::ctx::SpanContext;
@ -192,8 +192,8 @@ mod tests {
use arrow::record_batch::RecordBatch;
use arrow_util::{assert_batches_eq, assert_batches_sorted_eq};
use data_types::ColumnType;
use iox_query::frontend::sql::SqlQueryPlanner;
use iox_tests::util::TestCatalog;
use query::frontend::sql::SqlQueryPlanner;
#[tokio::test]
async fn test_query() {

View File

@ -1,9 +1,9 @@
//! Ring buffer of queries that have been run with some brief information
use data_types::NamespaceId;
use iox_query::QueryText;
use iox_time::{Time, TimeProvider};
use parking_lot::Mutex;
use query::QueryText;
use std::{
collections::VecDeque,
sync::{atomic, Arc},

View File

@ -8,9 +8,9 @@ use crate::{
};
use backoff::{Backoff, BackoffConfig};
use data_types::TableId;
use iox_query::{provider::ChunkPruner, QueryChunk};
use observability_deps::tracing::debug;
use predicate::Predicate;
use query::{provider::ChunkPruner, QueryChunk};
use schema::Schema;
use snafu::{ResultExt, Snafu};
use std::{collections::HashMap, sync::Arc};

View File

@ -9,12 +9,12 @@ use datafusion::{
logical_plan::Expr,
physical_plan::ExecutionPlan,
};
use predicate::{Predicate, PredicateBuilder};
use query::{
use iox_query::{
provider::{ChunkPruner, ProviderBuilder},
pruning::{prune_chunks, PruningObserver},
QueryChunk,
};
use predicate::{Predicate, PredicateBuilder};
use schema::Schema;
use super::QuerierTable;

View File

@ -25,7 +25,7 @@ once_cell = { version = "1.10.0", features = ["parking_lot"] }
predicate = { path = "../predicate" }
schema = { path = "../schema" }
querier = { path = "../querier" }
query = { path = "../query" }
iox_query = { path = "../iox_query" }
pretty_assertions = "1.2.1"
workspace-hack = { path = "../workspace-hack"}

View File

@ -1,8 +1,8 @@
use std::{any::Any, sync::Arc};
use datafusion::catalog::catalog::CatalogProvider;
use iox_query::{exec::ExecutionContextProvider, QueryDatabase};
use querier::QuerierNamespace;
use query::{exec::ExecutionContextProvider, QueryDatabase};
/// Abstract database used during testing.
pub trait AbstractDb: CatalogProvider + ExecutionContextProvider + QueryDatabase {

View File

@ -2,11 +2,11 @@ use crate::scenarios::*;
use arrow::datatypes::DataType;
use data_types::{MAX_NANO_TIME, MIN_NANO_TIME};
use datafusion::logical_plan::{col, lit};
use predicate::{rpc_predicate::InfluxRpcPredicate, PredicateBuilder};
use query::{
use iox_query::{
exec::fieldlist::{Field, FieldList},
frontend::influxrpc::InfluxRpcPlanner,
};
use predicate::{rpc_predicate::InfluxRpcPredicate, PredicateBuilder};
/// Creates and loads several database scenarios using the db_setup
/// function.

View File

@ -16,9 +16,9 @@ use crate::{
},
};
use datafusion::logical_plan::{col, lit, when};
use iox_query::frontend::influxrpc::InfluxRpcPlanner;
use predicate::rpc_predicate::InfluxRpcPredicate;
use predicate::PredicateBuilder;
use query::frontend::influxrpc::InfluxRpcPlanner;
use test_helpers::assert_contains;
/// runs read_filter(predicate) and compares it to the expected

View File

@ -14,9 +14,9 @@ use datafusion::{
logical_plan::{binary_expr, Operator},
prelude::*,
};
use iox_query::{frontend::influxrpc::InfluxRpcPlanner, Aggregate};
use predicate::rpc_predicate::InfluxRpcPredicate;
use predicate::PredicateBuilder;
use query::{frontend::influxrpc::InfluxRpcPlanner, Aggregate};
/// runs read_group(predicate) and compares it to the expected
/// output

View File

@ -1,9 +1,9 @@
//! Tests for the Influx gRPC queries
use crate::{influxrpc::util::run_series_set_plan, scenarios::*};
use datafusion::prelude::*;
use iox_query::{frontend::influxrpc::InfluxRpcPlanner, Aggregate, WindowDuration};
use predicate::rpc_predicate::InfluxRpcPredicate;
use predicate::PredicateBuilder;
use query::{frontend::influxrpc::InfluxRpcPlanner, Aggregate, WindowDuration};
/// runs read_window_aggregate(predicate) and compares it to the expected
/// output

View File

@ -2,11 +2,11 @@
use crate::scenarios::*;
use data_types::{MAX_NANO_TIME, MIN_NANO_TIME};
use datafusion::logical_plan::{col, lit};
use predicate::{rpc_predicate::InfluxRpcPredicate, PredicateBuilder};
use query::{
use iox_query::{
exec::stringset::{IntoStringSet, StringSetRef},
frontend::influxrpc::InfluxRpcPlanner,
};
use predicate::{rpc_predicate::InfluxRpcPredicate, PredicateBuilder};
/// runs table_names(predicate) and compares it to the expected
/// output

View File

@ -1,11 +1,11 @@
use crate::scenarios::*;
use data_types::{MAX_NANO_TIME, MIN_NANO_TIME};
use datafusion::logical_plan::{col, lit};
use predicate::{rpc_predicate::InfluxRpcPredicate, PredicateBuilder};
use query::{
use iox_query::{
exec::stringset::{IntoStringSet, StringSetRef},
frontend::influxrpc::InfluxRpcPlanner,
};
use predicate::{rpc_predicate::InfluxRpcPredicate, PredicateBuilder};
/// Creates and loads several database scenarios using the db_setup function.
///

View File

@ -1,10 +1,10 @@
use datafusion::logical_plan::{col, lit};
use predicate::rpc_predicate::InfluxRpcPredicate;
use predicate::PredicateBuilder;
use query::{
use iox_query::{
exec::stringset::{IntoStringSet, StringSetRef},
frontend::influxrpc::InfluxRpcPlanner,
};
use predicate::rpc_predicate::InfluxRpcPredicate;
use predicate::PredicateBuilder;
use crate::scenarios::*;

View File

@ -1,6 +1,6 @@
use datafusion::error::DataFusionError;
use query::exec::IOxSessionContext;
use query::plan::seriesset::SeriesSetPlans;
use iox_query::exec::IOxSessionContext;
use iox_query::plan::seriesset::SeriesSetPlans;
/// Run a series set plan to completion and produce a Vec<String> representation
///

View File

@ -5,7 +5,7 @@ mod setup;
use arrow::record_batch::RecordBatch;
use arrow_util::{display::pretty_format_batches, test_util::sort_record_batch};
use query::{
use iox_query::{
exec::{Executor, ExecutorType},
frontend::sql::SqlQueryPlanner,
};
@ -22,7 +22,7 @@ use self::{
setup::TestSetup,
};
use crate::scenarios::{DbScenario, DbSetup};
use query::exec::ExecutorConfig;
use iox_query::exec::ExecutorConfig;
#[allow(clippy::enum_variant_names)]
#[derive(Debug, Snafu)]

View File

@ -7,7 +7,7 @@ use super::{
use crate::scenarios::util::{make_n_chunks_scenario, ChunkData};
use async_trait::async_trait;
use data_types::{DeleteExpr, DeletePredicate, Op, Scalar, TimestampRange};
use query::frontend::sql::SqlQueryPlanner;
use iox_query::frontend::sql::SqlQueryPlanner;
#[derive(Debug)]
pub struct MeasurementWithMaxTime {}

View File

@ -9,7 +9,7 @@ use super::scenarios::*;
use arrow::record_batch::RecordBatch;
use arrow_util::assert_batches_sorted_eq;
use datafusion::error::DataFusionError;
use query::frontend::sql::SqlQueryPlanner;
use iox_query::frontend::sql::SqlQueryPlanner;
use test_helpers::assert_contains;
/// Runs the query in `sql` and compares it to the expected output.

View File

@ -1,7 +1,7 @@
//! Tests for the table_names implementation
use arrow::datatypes::DataType;
use query::QueryChunk;
use iox_query::QueryChunk;
use schema::selection::Selection;
use schema::{builder::SchemaBuilder, sort::SortKey, Schema, TIME_COLUMN_NAME};

View File

@ -9,7 +9,7 @@ edition = "2021"
# Workspace dependencies, in alphabetical order
datafusion = { path = "../datafusion" }
predicate = { path = "../predicate" }
query = { path = "../query" }
iox_query = { path = "../iox_query" }
workspace-hack = { path = "../workspace-hack"}
# Crates.io dependencies, in alphabetical order

View File

@ -5,7 +5,7 @@ pub mod planner;
use std::sync::Arc;
use async_trait::async_trait;
use query::{exec::ExecutionContextProvider, QueryDatabase};
use iox_query::{exec::ExecutionContextProvider, QueryDatabase};
/// Trait that allows the query engine (which includes flight and storage/InfluxRPC) to access a virtual set of
/// databases.

View File

@ -2,7 +2,7 @@
use std::sync::Arc;
use datafusion::physical_plan::ExecutionPlan;
use query::{
use iox_query::{
exec::IOxSessionContext,
frontend::{influxrpc::InfluxRpcPlanner, sql::SqlQueryPlanner},
plan::{fieldlist::FieldListPlan, seriesset::SeriesSetPlans, stringset::StringSetPlan},

View File

@ -11,7 +11,7 @@ data_types = { path = "../data_types" }
datafusion = { path = "../datafusion" }
generated_types = { path = "../generated_types" }
observability_deps = { path = "../observability_deps" }
query = { path = "../query" }
iox_query = { path = "../iox_query" }
service_common = { path = "../service_common" }
# Crates.io dependencies, in alphabetical order

View File

@ -16,13 +16,13 @@ use data_types::{DatabaseName, DatabaseNameError};
use datafusion::physical_plan::ExecutionPlan;
use futures::{SinkExt, Stream, StreamExt};
use generated_types::influxdata::iox::querier::v1 as proto;
use observability_deps::tracing::{info, warn};
use pin_project::{pin_project, pinned_drop};
use prost::Message;
use query::{
use iox_query::{
exec::{ExecutionContextProvider, IOxSessionContext},
QueryCompletedToken, QueryDatabase,
};
use observability_deps::tracing::{info, warn};
use pin_project::{pin_project, pinned_drop};
use prost::Message;
use serde::Deserialize;
use service_common::{planner::Planner, QueryDatabaseProvider};
use snafu::{ResultExt, Snafu};

View File

@ -10,7 +10,7 @@ datafusion = { path = "../datafusion" }
generated_types = { path = "../generated_types" }
observability_deps = { path = "../observability_deps" }
predicate = { path = "../predicate" }
query = { path = "../query" }
iox_query = { path = "../iox_query" }
query_functions = { path = "../query_functions"}
schema = { path = "../schema" }
service_common = { path = "../service_common" }

View File

@ -5,12 +5,12 @@ use std::{collections::BTreeSet, fmt, sync::Arc};
use arrow::datatypes::DataType as ArrowDataType;
use observability_deps::tracing::trace;
use predicate::rpc_predicate::{FIELD_COLUMN_NAME, MEASUREMENT_COLUMN_NAME};
use query::exec::{
use iox_query::exec::{
fieldlist::FieldList,
seriesset::series::{self, Either},
};
use observability_deps::tracing::trace;
use predicate::rpc_predicate::{FIELD_COLUMN_NAME, MEASUREMENT_COLUMN_NAME};
use generated_types::{
measurement_fields_response::{FieldType, MessageField},
@ -28,7 +28,7 @@ use snafu::Snafu;
pub enum Error {
#[snafu(display("Error converting series set to gRPC: {}", source))]
ConvertingSeries {
source: query::exec::seriesset::series::Error,
source: iox_query::exec::seriesset::series::Error,
},
#[snafu(display("Unsupported field data type in gRPC data translation: {}", data_type))]
@ -324,7 +324,7 @@ mod tests {
datatypes::DataType as ArrowDataType,
record_batch::RecordBatch,
};
use query::exec::{
use iox_query::exec::{
field::FieldIndexes,
fieldlist::Field,
seriesset::{

View File

@ -23,12 +23,12 @@ use generated_types::{
};
use super::{TAG_KEY_FIELD, TAG_KEY_MEASUREMENT};
use iox_query::{Aggregate as QueryAggregate, WindowDuration};
use observability_deps::tracing::warn;
use predicate::{
rpc_predicate::{InfluxRpcPredicate, FIELD_COLUMN_NAME, MEASUREMENT_COLUMN_NAME},
PredicateBuilder,
};
use query::{Aggregate as QueryAggregate, WindowDuration};
use snafu::{OptionExt, ResultExt, Snafu};
#[derive(Debug, Snafu)]

View File

@ -23,14 +23,14 @@ use generated_types::{
TagValuesGroupedByMeasurementAndTagKeyRequest, TagValuesRequest, TagValuesResponse,
TimestampRange,
};
use observability_deps::tracing::{error, info, trace};
use query::{
use iox_query::{
exec::{
fieldlist::FieldList, seriesset::converter::Error as SeriesSetError,
ExecutionContextProvider, IOxSessionContext,
},
QueryDatabase, QueryText,
};
use observability_deps::tracing::{error, info, trace};
use service_common::{planner::Planner, QueryDatabaseProvider};
use snafu::{OptionExt, ResultExt, Snafu};
use std::{
@ -1398,14 +1398,14 @@ mod tests {
generated_types::*,
Client as StorageClient, OrgAndBucket,
};
use iox_query::{
exec::Executor,
test::{TestChunk, TestDatabase},
};
use metric::{Attributes, Metric, U64Counter};
use panic_logging::SendPanicsToTracing;
use parking_lot::Mutex;
use predicate::{PredicateBuilder, PredicateMatch};
use query::{
exec::Executor,
test::{TestChunk, TestDatabase},
};
use service_common::QueryDatabaseProvider;
use std::{
collections::BTreeMap,