refactor: Rename more mentions of Kafka to their other name where appropriate

pull/24376/head
Carol (Nichols || Goulding) 2022-08-25 16:56:14 -04:00
parent 1b49ad25f7
commit dbd27f648f
No known key found for this signature in database
GPG Key ID: E907EE5A736F87D4
9 changed files with 49 additions and 49 deletions

View File

@ -40,7 +40,7 @@ enum Command {
/// Run database migrations
Setup(Setup),
/// Manage kafka topic
/// Manage topic
Topic(topic::Config),
}

View File

@ -83,7 +83,7 @@ fn shared_handle(handle: JoinHandle<()>) -> SharedJoinHandle {
/// persistence and answer queries
#[derive(Debug)]
pub struct IngestHandlerImpl<T = SystemProvider> {
/// Kafka Topic assigned to this ingester
/// Topic assigned to this ingester
#[allow(dead_code)]
topic: TopicMetadata,

View File

@ -129,8 +129,8 @@ pub struct LifecycleManager {
/// The configuration options for the lifecycle on the ingester.
#[derive(Debug, Clone, Copy)]
pub struct LifecycleConfig {
/// The ingester will pause pulling data from Kafka if it hits this amount of memory used, waiting
/// until persistence evicts partitions from memory.
/// The ingester will pause pulling data from the write buffer if it hits this amount of memory
/// used, waiting until persistence evicts partitions from memory.
pause_ingest_size: usize,
/// When the ingester hits this threshold, the lifecycle manager will persist the largest
/// partitions currently buffered until it falls below this threshold. An ingester running
@ -147,7 +147,7 @@ pub struct LifecycleConfig {
partition_size_threshold: usize,
/// If an individual partitiion has had data buffered for longer than this period of time, the
/// manager will persist it. This setting is to ensure we have an upper bound on how far back
/// we will need to read in Kafka on restart or recovery.
/// we will need to read in the write buffer on restart or recovery.
partition_age_threshold: Duration,
/// If an individual partition hasn't received a write for longer than this period of time, the
/// manager will persist it. This is to ensure that cold partitions get cleared out to make

View File

@ -64,7 +64,7 @@ pub struct SequencedStreamHandler<I, O, T = SystemProvider> {
shard_reset_count: U64Counter,
/// Log context fields - otherwise unused.
kafka_topic_name: String,
topic_name: String,
shard_index: ShardIndex,
skip_to_oldest_available: bool,
@ -83,7 +83,7 @@ impl<I, O> SequencedStreamHandler<I, O> {
current_sequence_number: SequenceNumber,
sink: O,
lifecycle_handle: LifecycleHandleImpl,
kafka_topic_name: String,
topic_name: String,
shard_index: ShardIndex,
metrics: &metric::Registry,
skip_to_oldest_available: bool,
@ -92,7 +92,7 @@ impl<I, O> SequencedStreamHandler<I, O> {
let time_to_be_readable = metrics.register_metric::<DurationGauge>(
"ingester_ttbr",
"duration of time between producer writing to consumer putting into queryable cache",
).recorder(metric_attrs(shard_index, &kafka_topic_name, None, false));
).recorder(metric_attrs(shard_index, &topic_name, None, false));
// Lifecycle-driven ingest pause duration
let pause_duration = metrics
@ -109,31 +109,31 @@ impl<I, O> SequencedStreamHandler<I, O> {
);
let shard_unknown_sequence_number_count = ingest_errors.recorder(metric_attrs(
shard_index,
&kafka_topic_name,
&topic_name,
Some("shard_unknown_sequence_number"),
true,
));
let shard_invalid_data_count = ingest_errors.recorder(metric_attrs(
shard_index,
&kafka_topic_name,
&topic_name,
Some("shard_invalid_data"),
true,
));
let shard_unknown_error_count = ingest_errors.recorder(metric_attrs(
shard_index,
&kafka_topic_name,
&topic_name,
Some("shard_unknown_error"),
true,
));
let sink_apply_error_count = ingest_errors.recorder(metric_attrs(
shard_index,
&kafka_topic_name,
&topic_name,
Some("sink_apply_error"),
true,
));
let skipped_sequence_number_amount = ingest_errors.recorder(metric_attrs(
shard_index,
&kafka_topic_name,
&topic_name,
Some("skipped_sequence_number_amount"),
true,
));
@ -144,7 +144,7 @@ impl<I, O> SequencedStreamHandler<I, O> {
"shard_reset_count",
"how often a shard was already reset",
)
.recorder(metric_attrs(shard_index, &kafka_topic_name, None, true));
.recorder(metric_attrs(shard_index, &topic_name, None, true));
Self {
write_buffer_stream_handler,
@ -160,7 +160,7 @@ impl<I, O> SequencedStreamHandler<I, O> {
sink_apply_error_count,
skipped_sequence_number_amount,
shard_reset_count,
kafka_topic_name,
topic_name,
shard_index,
skip_to_oldest_available,
}
@ -183,7 +183,7 @@ impl<I, O> SequencedStreamHandler<I, O> {
sink_apply_error_count: self.sink_apply_error_count,
skipped_sequence_number_amount: self.skipped_sequence_number_amount,
shard_reset_count: self.shard_reset_count,
kafka_topic_name: self.kafka_topic_name,
topic_name: self.topic_name,
shard_index: self.shard_index,
skip_to_oldest_available: self.skip_to_oldest_available,
}
@ -220,7 +220,7 @@ where
next = stream.next().fuse() => next,
_ = shutdown_fut => {
info!(
kafka_topic=%self.kafka_topic_name,
kafka_topic=%self.topic_name,
shard_index=%self.shard_index,
"stream handler shutdown",
);
@ -260,7 +260,7 @@ where
if self.skip_to_oldest_available && sequence_number_before_reset.is_none() {
warn!(
error=%e,
kafka_topic=%self.kafka_topic_name,
kafka_topic=%self.topic_name,
shard_index=%self.shard_index,
potential_data_loss=true,
"reset stream"
@ -273,7 +273,7 @@ where
} else {
error!(
error=%e,
kafka_topic=%self.kafka_topic_name,
kafka_topic=%self.topic_name,
shard_index=%self.shard_index,
potential_data_loss=true,
"unable to read from desired sequence number offset"
@ -285,7 +285,7 @@ where
Some(Err(e)) if e.kind() == WriteBufferErrorKind::IO => {
warn!(
error=%e,
kafka_topic=%self.kafka_topic_name,
kafka_topic=%self.topic_name,
shard_index=%self.shard_index,
"I/O error reading from shard"
);
@ -300,7 +300,7 @@ where
// be applied/persisted.
error!(
error=%e,
kafka_topic=%self.kafka_topic_name,
kafka_topic=%self.topic_name,
shard_index=%self.shard_index,
potential_data_loss=true,
"unable to deserialize dml operation"
@ -316,13 +316,13 @@ Shard Index {:?} stream for topic {} has a high watermark BEFORE the sequence nu
is either a bug (see https://github.com/influxdata/rskafka/issues/147 for example) or means that \
someone re-created the shard and data is lost. In both cases, it's better to panic than to try \
something clever.",
self.shard_index, self.kafka_topic_name,
self.shard_index, self.topic_name,
)
}
Some(Err(e)) => {
error!(
error=%e,
kafka_topic=%self.kafka_topic_name,
kafka_topic=%self.topic_name,
shard_index=%self.shard_index,
potential_data_loss=true,
"unhandled error converting write buffer data to DmlOperation",
@ -334,7 +334,7 @@ something clever.",
None => {
panic!(
"shard index {:?} stream for topic {} ended without graceful shutdown",
self.shard_index, self.kafka_topic_name
self.shard_index, self.topic_name
);
}
};
@ -349,7 +349,7 @@ something clever.",
if let Some(op) = op {
// Emit per-op debug info.
trace!(
kafka_topic=%self.kafka_topic_name,
kafka_topic=%self.topic_name,
shard_index=%self.shard_index,
op_size=op.size(),
op_namespace=op.namespace(),
@ -366,7 +366,7 @@ something clever.",
let should_pause = match self.sink.apply(op).await {
Ok(should_pause) => {
trace!(
kafka_topic=%self.kafka_topic_name,
kafka_topic=%self.topic_name,
shard_index=%self.shard_index,
%should_pause,
"successfully applied dml operation"
@ -376,7 +376,7 @@ something clever.",
Err(e) => {
error!(
error=%e,
kafka_topic=%self.kafka_topic_name,
kafka_topic=%self.topic_name,
shard_index=%self.shard_index,
potential_data_loss=true,
"failed to apply dml operation"
@ -404,7 +404,7 @@ something clever.",
let started_at = self.time_provider.now();
warn!(
kafka_topic=%self.kafka_topic_name,
kafka_topic=%self.topic_name,
shard_index=%self.shard_index,
"pausing ingest until persistence has run"
);
@ -429,7 +429,7 @@ something clever.",
.unwrap_or_else(|| "unknown".to_string());
info!(
kafka_topic=%self.kafka_topic_name,
kafka_topic=%self.topic_name,
shard_index=%self.shard_index,
pause_duration=%duration_str,
"resuming ingest"
@ -483,7 +483,7 @@ mod tests {
static TEST_TIME: Lazy<Time> = Lazy::new(|| SystemProvider::default().now());
static TEST_SHARD_INDEX: ShardIndex = ShardIndex::new(42);
static TEST_KAFKA_TOPIC: &str = "kafka_topic_name";
static TEST_TOPIC_NAME: &str = "topic_name";
// Return a DmlWrite with the given namespace and a single table.
fn make_write(name: impl Into<String>, write_time: u64) -> DmlWrite {
@ -633,7 +633,7 @@ mod tests {
SequenceNumber::new(0),
Arc::clone(&sink),
lifecycle.handle(),
TEST_KAFKA_TOPIC.to_string(),
TEST_TOPIC_NAME.to_string(),
TEST_SHARD_INDEX,
&*metrics,
$skip_to_oldest_available,
@ -678,7 +678,7 @@ mod tests {
.get_instrument::<Metric<DurationGauge>>("ingester_ttbr")
.expect("did not find ttbr metric")
.get_observer(&Attributes::from([
("kafka_topic", TEST_KAFKA_TOPIC.into()),
("kafka_topic", TEST_TOPIC_NAME.into()),
("kafka_partition", TEST_SHARD_INDEX.to_string().into()),
]))
.expect("did not match metric attributes")
@ -690,7 +690,7 @@ mod tests {
.get_instrument::<Metric<U64Counter>>("shard_reset_count")
.expect("did not find reset count metric")
.get_observer(&Attributes::from([
("kafka_topic", TEST_KAFKA_TOPIC.into()),
("kafka_topic", TEST_TOPIC_NAME.into()),
("kafka_partition", TEST_SHARD_INDEX.to_string().into()),
("potential_data_loss", "true".into()),
]))
@ -705,7 +705,7 @@ mod tests {
.expect("did not find error metric")
.get_observer(&metric_attrs(
TEST_SHARD_INDEX,
TEST_KAFKA_TOPIC,
TEST_TOPIC_NAME,
Some($metric_name),
true,
))
@ -967,7 +967,7 @@ mod tests {
// An abnormal end to the steam causes a panic, rather than a silent stream reader exit.
#[tokio::test]
#[should_panic(
expected = "shard index ShardIndex(42) stream for topic kafka_topic_name ended without \
expected = "shard index ShardIndex(42) stream for topic topic_name ended without \
graceful shutdown"
)]
async fn test_early_stream_end_panic() {
@ -988,7 +988,7 @@ mod tests {
SequenceNumber::new(0),
sink,
lifecycle.handle(),
"kafka_topic_name".to_string(),
"topic_name".to_string(),
ShardIndex::new(42),
&*metrics,
false,
@ -1028,7 +1028,7 @@ mod tests {
SequenceNumber::new(0),
sink,
lifecycle.handle(),
"kafka_topic_name".to_string(),
"topic_name".to_string(),
ShardIndex::new(42),
&*metrics,
false,

View File

@ -79,13 +79,13 @@ where
pub fn new(
inner: T,
watermark_fetcher: F,
kafka_topic_name: String,
topic_name: String,
shard_index: ShardIndex,
metrics: &metric::Registry,
) -> Self {
let attr = Attributes::from([
("kafka_partition", shard_index.to_string().into()),
("kafka_topic", kafka_topic_name.into()),
("kafka_topic", topic_name.into()),
]);
let write_buffer_bytes_read = metrics
@ -250,16 +250,16 @@ mod tests {
/// be observing for.
const SHARD_INDEX: ShardIndex = ShardIndex::new(42);
static TEST_KAFKA_TOPIC: &str = "kafka_topic_name";
static TEST_TOPIC_NAME: &str = "topic_name";
static TEST_TIME: Lazy<Time> = Lazy::new(|| SystemProvider::default().now());
/// The attributes assigned to the metrics emitted by the
/// instrumentation when using the above shard / kafka topic values.
/// instrumentation when using the above shard / topic values.
static DEFAULT_ATTRS: Lazy<Attributes> = Lazy::new(|| {
Attributes::from([
("kafka_partition", SHARD_INDEX.to_string().into()),
("kafka_topic", TEST_KAFKA_TOPIC.into()),
("kafka_topic", TEST_TOPIC_NAME.into()),
])
});
@ -296,7 +296,7 @@ mod tests {
let instrumentation = SinkInstrumentation::new(
inner,
MockWatermarkFetcher::new(with_fetcher_return),
TEST_KAFKA_TOPIC.to_string(),
TEST_TOPIC_NAME.to_string(),
SHARD_INDEX,
metrics,
);

View File

@ -19,8 +19,8 @@ use mutable_batch::MutableBatch;
use std::{borrow::Cow, collections::BTreeMap};
use thiserror::Error;
const SHARED_KAFKA_TOPIC: &str = "iox-shared";
const SHARED_QUERY_POOL: &str = SHARED_KAFKA_TOPIC;
const SHARED_TOPIC_NAME: &str = "iox-shared";
const SHARED_QUERY_POOL: &str = SHARED_TOPIC_NAME;
const TIME_COLUMN: &str = "time";
/// A string value representing an infinite retention policy.
@ -202,7 +202,7 @@ pub async fn create_or_get_default_records(
shard_count: i32,
txn: &mut dyn Transaction,
) -> Result<(TopicMetadata, QueryPool, BTreeMap<ShardId, Shard>)> {
let topic = txn.topics().create_or_get(SHARED_KAFKA_TOPIC).await?;
let topic = txn.topics().create_or_get(SHARED_TOPIC_NAME).await?;
let query_pool = txn.query_pools().create_or_get(SHARED_QUERY_POOL).await?;
let mut shards = BTreeMap::new();

View File

@ -155,7 +155,7 @@ pub enum Error {
#[error("failed to initialise write buffer connection: {0}")]
WriteBuffer(#[from] write_buffer::core::WriteBufferError),
#[error("failed to create KafkaPartition from id: {0}")]
#[error("failed to create ShardIndex from id: {0}")]
InvalidData(#[from] std::num::TryFromIntError),
#[error("querier error: {0}")]

View File

@ -8,7 +8,7 @@
//! * Enforcing schema validation & synchronising it within the catalog.
//! * Deriving the partition key of each DML operation.
//! * Applying sharding logic.
//! * Push resulting operations into the appropriate kafka partitions.
//! * Push resulting operations into the appropriate shards (Kafka partitions if using Kafka).
#![deny(
rustdoc::broken_intra_doc_links,

View File

@ -37,7 +37,7 @@ pub async fn initialize_db(dsn: &str, schema_name: &str) {
.ok()
.unwrap();
// Create the shared Kafka topic in the catalog
// Create the shared topic in the catalog
Command::cargo_bin("influxdb_iox")
.unwrap()
.arg("catalog")