diff --git a/src/common/config/src/wal/kafka.rs b/src/common/config/src/wal/kafka.rs index e93aa6cb2271..858991264bb6 100644 --- a/src/common/config/src/wal/kafka.rs +++ b/src/common/config/src/wal/kafka.rs @@ -42,7 +42,7 @@ pub struct KafkaConfig { #[serde(skip)] #[serde(default)] pub compression: RsKafkaCompression, - /// The maximum log size a kakfa batch producer could buffer. + /// The maximum log size a kafka batch producer could buffer. pub max_batch_size: ReadableSize, /// The linger duration of a kafka batch producer. #[serde(with = "humantime_serde")] diff --git a/src/mito2/src/sst/index/applier.rs b/src/mito2/src/sst/index/applier.rs index 564a20d33c0f..633aad007f1a 100644 --- a/src/mito2/src/sst/index/applier.rs +++ b/src/mito2/src/sst/index/applier.rs @@ -28,7 +28,7 @@ pub struct SstIndexApplier { /// Object store responsible for accessing SST files. object_store: ObjectStore, - /// Predifined index applier used to apply predicates to index files + /// Predefined index applier used to apply predicates to index files /// and return the relevant row group ids for further scan. index_applier: Arc, } diff --git a/src/mito2/src/sst/index/applier/builder.rs b/src/mito2/src/sst/index/applier/builder.rs index 6ecd0b0e5d5c..6414081ac713 100644 --- a/src/mito2/src/sst/index/applier/builder.rs +++ b/src/mito2/src/sst/index/applier/builder.rs @@ -14,7 +14,7 @@ mod between; -// TODO(zhongzc): This PR is too large. The following modules are comming soon. +// TODO(zhongzc): This PR is too large. The following modules are coming soon. // mod comparison; // mod eq_list; @@ -99,7 +99,7 @@ impl<'a> SstIndexApplierBuilder<'a> { match expr { DfExpr::Between(between) => self.collect_between(between), - // TODO(zhongzc): This PR is too large. The following arms are comming soon. + // TODO(zhongzc): This PR is too large. The following arms are coming soon. // DfExpr::InList(in_list) => self.collect_inlist(in_list), // DfExpr::BinaryExpr(BinaryExpr { left, op, right }) => match op { @@ -145,7 +145,7 @@ impl<'a> SstIndexApplierBuilder<'a> { .then(|| column.column_schema.data_type.clone())) } - /// Helper funtion to get a non-null literal. + /// Helper function to get a non-null literal. fn nonnull_lit(expr: &DfExpr) -> Option<&ScalarValue> { match expr { DfExpr::Literal(lit) if !lit.is_null() => Some(lit),