Merge branch 'main' into dependabot/cargo/arrow-5.5.0
commit
181145eca1
|
|
@ -2658,9 +2658,9 @@ dependencies = [
|
|||
|
||||
[[package]]
|
||||
name = "parquet"
|
||||
version = "5.4.0"
|
||||
version = "5.5.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "f231efd0363ac0d8e6e98113e1913b8298ec15fccb42f142c386831fccbe9ddd"
|
||||
checksum = "a01fdb20e85f8403bb3939f79640dc606f104c7e5d1c8235444ccc798f466ac8"
|
||||
dependencies = [
|
||||
"arrow",
|
||||
"base64 0.13.0",
|
||||
|
|
|
|||
|
|
@ -132,7 +132,7 @@ log = "0.4"
|
|||
once_cell = { version = "1.4.0", features = ["parking_lot"] }
|
||||
parking_lot = "0.11.2"
|
||||
itertools = "0.10.1"
|
||||
parquet = "5.3"
|
||||
parquet = "5.5"
|
||||
# used by arrow/datafusion anyway
|
||||
prettytable-rs = "0.8"
|
||||
pprof = { version = "^0.5", default-features = false, features = ["flamegraph", "protobuf"], optional = true }
|
||||
|
|
|
|||
|
|
@ -10,7 +10,7 @@ influxdb_tsm = { path = "../influxdb_tsm" }
|
|||
internal_types = { path = "../internal_types" }
|
||||
snafu = "0.6.2"
|
||||
observability_deps = { path = "../observability_deps" }
|
||||
parquet = "5.3"
|
||||
parquet = "5.5"
|
||||
|
||||
[dev-dependencies] # In alphabetical order
|
||||
criterion = "0.3.3"
|
||||
|
|
|
|||
|
|
@ -313,7 +313,7 @@ File {
|
|||
},
|
||||
),
|
||||
file_size_bytes: 33,
|
||||
metadata: b"metadata omitted (930 bytes)",
|
||||
metadata: b"metadata omitted (933 bytes)",
|
||||
},
|
||||
),
|
||||
),
|
||||
|
|
@ -437,7 +437,7 @@ File {
|
|||
},
|
||||
),
|
||||
file_size_bytes: 33,
|
||||
metadata: b"metadata omitted (930 bytes)",
|
||||
metadata: b"metadata omitted (933 bytes)",
|
||||
},
|
||||
),
|
||||
),
|
||||
|
|
|
|||
|
|
@ -1772,7 +1772,7 @@ mod tests {
|
|||
.id();
|
||||
|
||||
// A chunk is now in the object store and still in read buffer
|
||||
let expected_parquet_size = 1244;
|
||||
let expected_parquet_size = 1243;
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "read_buffer", expected_read_buffer_size);
|
||||
// now also in OS
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "object_store", expected_parquet_size);
|
||||
|
|
@ -2211,7 +2211,7 @@ mod tests {
|
|||
// Read buffer + Parquet chunk size
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "mutable_buffer", 0);
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "read_buffer", 1700);
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "object_store", 1242);
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "object_store", 1241);
|
||||
|
||||
// All the chunks should have different IDs
|
||||
assert_ne!(mb_chunk.id(), rb_chunk.id());
|
||||
|
|
@ -2326,7 +2326,7 @@ mod tests {
|
|||
// Read buffer + Parquet chunk size
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "mutable_buffer", 0);
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "read_buffer", 1700);
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "object_store", 1242);
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "object_store", 1241);
|
||||
|
||||
// Unload RB chunk but keep it in OS
|
||||
let pq_chunk = db
|
||||
|
|
@ -2347,7 +2347,7 @@ mod tests {
|
|||
// Parquet chunk size only
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "mutable_buffer", 0);
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "read_buffer", 0);
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "object_store", 1242);
|
||||
catalog_chunk_size_bytes_metric_eq(registry, "object_store", 1241);
|
||||
|
||||
// Verify data written to the parquet file in object store
|
||||
//
|
||||
|
|
|
|||
Loading…
Reference in New Issue