influxdb/compactor
Marco Neumann e0062f2d40
refactor: do NOT use fake DF context for parquet reading (#5942)
Use the proper top-level DataFusion context and register the object
store there.

Note that we still hide the `ParquetExec` behind an opaque record batch
stream. Fixing that is next on my list.

Helps with #5897.

Co-authored-by: kodiakhq[bot] <49736102+kodiakhq[bot]@users.noreply.github.com>
2022-10-24 08:20:26 +00:00
..
src refactor: do NOT use fake DF context for parquet reading (#5942) 2022-10-24 08:20:26 +00:00
Cargo.toml feat: Create compactor service to list skipped compactions 2022-10-21 13:40:31 -04:00
README.md docs: add consensus for the desired final output of the compactor (#5069) 2022-07-07 19:11:16 +00:00

README.md

After a partition of a table has not received any writes for some amount of time, the compactor will ensure it is stored in object store as N parquet files which:

  • have non overlapping time ranges
  • each does not exceed a size specified by config param max_desired_file_size_bytes.