From 53b56d71262c48e189e22bfcf652b31b05b692da Mon Sep 17 00:00:00 2001 From: yew1eb Date: Sun, 11 Jan 2026 17:58:32 +0800 Subject: [PATCH] [AURON #1877] Fix spelling errors via the typos tool --- native-engine/auron-jni-bridge/src/jni_bridge.rs | 2 +- native-engine/auron-serde/src/from_proto.rs | 4 ++-- native-engine/datafusion-ext-commons/src/arrow/selection.rs | 2 +- .../datafusion-ext-commons/src/io/ipc_compression.rs | 2 +- native-engine/datafusion-ext-exprs/src/spark_udf_wrapper.rs | 2 +- native-engine/datafusion-ext-exprs/src/string_ends_with.rs | 4 ++-- .../datafusion-ext-plans/src/common/cached_exprs_evaluator.rs | 2 +- native-engine/datafusion-ext-plans/src/sort_exec.rs | 2 +- .../spark/sql/execution/auron/plan/NativeWindowBase.scala | 2 +- 9 files changed, 11 insertions(+), 11 deletions(-) diff --git a/native-engine/auron-jni-bridge/src/jni_bridge.rs b/native-engine/auron-jni-bridge/src/jni_bridge.rs index 90614cc23..6ea65ca85 100644 --- a/native-engine/auron-jni-bridge/src/jni_bridge.rs +++ b/native-engine/auron-jni-bridge/src/jni_bridge.rs @@ -540,7 +540,7 @@ impl JavaClasses<'static> { pub fn get() -> &'static JavaClasses<'static> { unsafe { - // safety: JNI_JAVA_CLASSES must be initialized frist + // safety: JNI_JAVA_CLASSES must be initialized first JNI_JAVA_CLASSES.get_unchecked() } } diff --git a/native-engine/auron-serde/src/from_proto.rs b/native-engine/auron-serde/src/from_proto.rs index 0caaad6ca..09c700db0 100644 --- a/native-engine/auron-serde/src/from_proto.rs +++ b/native-engine/auron-serde/src/from_proto.rs @@ -1075,7 +1075,7 @@ fn try_parse_physical_sort_expr( input: &Arc, sort: &Box, ) -> Result, PlanSerDeError> { - let pyhsical_sort_expr = sort + let physical_sort_expr = sort .expr .iter() .map(|expr| { @@ -1111,7 +1111,7 @@ fn try_parse_physical_sort_expr( } }) .collect::, _>>()?; - Ok(pyhsical_sort_expr) + Ok(physical_sort_expr) } pub fn parse_protobuf_partitioning( diff --git a/native-engine/datafusion-ext-commons/src/arrow/selection.rs b/native-engine/datafusion-ext-commons/src/arrow/selection.rs index bf67bc28d..32699086c 100644 --- a/native-engine/datafusion-ext-commons/src/arrow/selection.rs +++ b/native-engine/datafusion-ext-commons/src/arrow/selection.rs @@ -216,7 +216,7 @@ pub fn create_array_interleaver( NonNull::new_unchecked(take_value_ptrs[take_value_ptr_idx] as *mut u8); take_value_ptr_idx += 1; - // for continous elements, just extend the area to copy + // for continuous elements, just extend the area to copy // otherwise, copy current area end move to the next area if src_end_ptr != value_ptr { prefetch_read_data!(value_ptr.as_ptr()); // prefetch next while copying current diff --git a/native-engine/datafusion-ext-commons/src/io/ipc_compression.rs b/native-engine/datafusion-ext-commons/src/io/ipc_compression.rs index d3ea0b293..817b2c8e2 100644 --- a/native-engine/datafusion-ext-commons/src/io/ipc_compression.rs +++ b/native-engine/datafusion-ext-commons/src/io/ipc_compression.rs @@ -204,7 +204,7 @@ impl IoCompressionWriter { match self { IoCompressionWriter::LZ4(w) => { w.try_finish() - .or_else(|_| df_execution_err!("ipc compresion error"))?; + .or_else(|_| df_execution_err!("ipc compression error"))?; } IoCompressionWriter::ZSTD(w) => { w.do_finish()?; diff --git a/native-engine/datafusion-ext-exprs/src/spark_udf_wrapper.rs b/native-engine/datafusion-ext-exprs/src/spark_udf_wrapper.rs index e3c43e6e6..075865ccf 100644 --- a/native-engine/datafusion-ext-exprs/src/spark_udf_wrapper.rs +++ b/native-engine/datafusion-ext-exprs/src/spark_udf_wrapper.rs @@ -209,7 +209,7 @@ fn invoke_udf( params_batch: RecordBatch, result_schema: SchemaRef, ) -> Result { - // evalute via context + // evaluate via context let struct_array = StructArray::from(params_batch); let mut export_ffi_array = FFI_ArrowArray::new(&struct_array.to_data()); let mut import_ffi_array = FFI_ArrowArray::empty(); diff --git a/native-engine/datafusion-ext-exprs/src/string_ends_with.rs b/native-engine/datafusion-ext-exprs/src/string_ends_with.rs index 1432a24cb..1d1fd9583 100644 --- a/native-engine/datafusion-ext-exprs/src/string_ends_with.rs +++ b/native-engine/datafusion-ext-exprs/src/string_ends_with.rs @@ -139,10 +139,10 @@ mod test { Some("rr".to_string()), Some("roser r".to_string()), ])); - // create a shema with the field + // create a schema with the field let schema = Arc::new(Schema::new(vec![Field::new("col2", DataType::Utf8, true)])); - // create a RecordBatch with the shema and StringArray + // create a RecordBatch with the schema and StringArray let batch = RecordBatch::try_new(schema, vec![string_array]).expect("Error creating RecordBatch"); diff --git a/native-engine/datafusion-ext-plans/src/common/cached_exprs_evaluator.rs b/native-engine/datafusion-ext-plans/src/common/cached_exprs_evaluator.rs index 1c7ec79fe..fe77a4382 100644 --- a/native-engine/datafusion-ext-plans/src/common/cached_exprs_evaluator.rs +++ b/native-engine/datafusion-ext-plans/src/common/cached_exprs_evaluator.rs @@ -402,7 +402,7 @@ impl PhysicalExpr for CachedExpr { } } -/// A struct holding all evaluated values of cachable expressions +/// A struct holding all evaluated values of cacheable expressions #[derive(Clone)] struct Cache { values: Arc>>>, diff --git a/native-engine/datafusion-ext-plans/src/sort_exec.rs b/native-engine/datafusion-ext-plans/src/sort_exec.rs index 24d0beb6b..8e20bc41c 100644 --- a/native-engine/datafusion-ext-plans/src/sort_exec.rs +++ b/native-engine/datafusion-ext-plans/src/sort_exec.rs @@ -1266,7 +1266,7 @@ impl KeyCollector for InMemRowsKeyCollector { assert_eq!(self.offsets.len() - 1, num_rows); let mut rows = row_converter.empty_rows(0, 0); unsafe { - // safety: acces rows.buffer/offsets + // safety: access rows.buffer/offsets struct XRows { buffer: Vec, offsets: Vec, diff --git a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeWindowBase.scala b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeWindowBase.scala index 2e8495368..a7a0792bc 100644 --- a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeWindowBase.scala +++ b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeWindowBase.scala @@ -212,7 +212,7 @@ abstract class NativeWindowBase( .addAllPartitionSpec(nativePartitionSpecExprs.asJava) .addAllOrderSpec(nativeOrderSpecExprs.asJava) - // WindowGrupLimitExec does not output window cols + // WindowGroupLimitExec does not output window cols groupLimit match { case Some(limit) => nativeWindowExec.setGroupLimit(WindowGroupLimit.newBuilder().setK(limit))