Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion benchmarks/src/bin/external_aggr.rs
Original file line number Diff line number Diff line change
Expand Up @@ -113,7 +113,7 @@ impl ExternalAggrConfig {
"#,
];

/// If `--query` and `--memory-limit` is not speicified, run all queries
/// If `--query` and `--memory-limit` is not specified, run all queries
/// with pre-configured memory limits
/// If only `--query` is specified, run the query with all memory limits
/// for this query
Expand Down
2 changes: 1 addition & 1 deletion datafusion-cli/tests/cli_integration.rs
Original file line number Diff line number Diff line change
Expand Up @@ -64,7 +64,7 @@ async fn setup_minio_container() -> ContainerAsync<minio::MinIO> {

match container {
Ok(container) => {
// We wait for MinIO to be healthy and preprare test files. We do it via CLI to avoid s3 dependency
// We wait for MinIO to be healthy and prepare test files. We do it via CLI to avoid s3 dependency
let commands = [
ExecCommand::new(["/usr/bin/mc", "ready", "local"]),
ExecCommand::new([
Expand Down
2 changes: 1 addition & 1 deletion datafusion-examples/examples/custom_file_casts.rs
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ use object_store::path::Path;
use object_store::{ObjectStore, PutPayload};

// Example showing how to implement custom casting rules to adapt file schemas.
// This example enforces that casts must be stricly widening: if the file type is Int64 and the table type is Int32, it will error
// This example enforces that casts must be strictly widening: if the file type is Int64 and the table type is Int32, it will error
// before even reading the data.
// Without this custom cast rule DataFusion would happily do the narrowing cast, potentially erroring only if it found a row with data it could not cast.

Expand Down
4 changes: 2 additions & 2 deletions datafusion-examples/examples/expr_api.rs
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@ async fn main() -> Result<()> {
boundary_analysis_and_selectivity_demo()?;

// See how boundary analysis works for `AND` & `OR` conjunctions.
boundary_analysis_in_conjuctions_demo()?;
boundary_analysis_in_conjunctions_demo()?;

// See how to determine the data types of expressions
expression_type_demo()?;
Expand Down Expand Up @@ -351,7 +351,7 @@ fn boundary_analysis_and_selectivity_demo() -> Result<()> {

/// This function shows how to think about and leverage the analysis API
/// to infer boundaries in `AND` & `OR` conjunctions.
fn boundary_analysis_in_conjuctions_demo() -> Result<()> {
fn boundary_analysis_in_conjunctions_demo() -> Result<()> {
// Let us consider the more common case of AND & OR conjunctions.
//
// age > 18 AND age <= 25
Expand Down
2 changes: 1 addition & 1 deletion datafusion/catalog/src/information_schema.rs
Original file line number Diff line number Diff line change
Expand Up @@ -810,7 +810,7 @@ impl InformationSchemaColumnsBuilder {
) {
use DataType::*;

// Note: append_value is actually infallable.
// Note: append_value is actually infallible.
self.catalog_names.append_value(catalog_name);
self.schema_names.append_value(schema_name);
self.table_names.append_value(table_name);
Expand Down
4 changes: 2 additions & 2 deletions datafusion/common/src/test_util.rs
Original file line number Diff line number Diff line change
Expand Up @@ -158,7 +158,7 @@ macro_rules! assert_batches_sorted_eq {
/// Is a macro so test error
/// messages are on the same line as the failure;
///
/// Both arguments must be convertable into Strings ([`Into`]<[`String`]>)
/// Both arguments must be convertible into Strings ([`Into`]<[`String`]>)
#[macro_export]
macro_rules! assert_contains {
($ACTUAL: expr, $EXPECTED: expr) => {
Expand All @@ -181,7 +181,7 @@ macro_rules! assert_contains {
/// Is a macro so test error
/// messages are on the same line as the failure;
///
/// Both arguments must be convertable into Strings ([`Into`]<[`String`]>)
/// Both arguments must be convertible into Strings ([`Into`]<[`String`]>)
#[macro_export]
macro_rules! assert_not_contains {
($ACTUAL: expr, $UNEXPECTED: expr) => {
Expand Down
4 changes: 2 additions & 2 deletions datafusion/core/benches/spm.rs
Original file line number Diff line number Diff line change
Expand Up @@ -66,7 +66,7 @@ fn generate_spm_for_round_robin_tie_breaker(
};

let rbs = (0..batch_count).map(|_| rb.clone()).collect::<Vec<_>>();
let partitiones = vec![rbs.clone(); partition_count];
let partitions = vec![rbs.clone(); partition_count];

let schema = rb.schema();
let sort = [
Expand All @@ -81,7 +81,7 @@ fn generate_spm_for_round_robin_tie_breaker(
]
.into();

let exec = MemorySourceConfig::try_new_exec(&partitiones, schema, None).unwrap();
let exec = MemorySourceConfig::try_new_exec(&partitions, schema, None).unwrap();
SortPreservingMergeExec::new(sort, exec)
.with_round_robin_repartition(enable_round_robin_repartition)
}
Expand Down
2 changes: 1 addition & 1 deletion datafusion/core/src/bin/print_functions_docs.rs
Original file line number Diff line number Diff line change
Expand Up @@ -260,7 +260,7 @@ fn print_docs(
}
}

/// Trait for accessing name / aliases / documentation for differnet functions
/// Trait for accessing name / aliases / documentation for different functions
trait DocProvider {
fn get_name(&self) -> String;
fn get_aliases(&self) -> Vec<String>;
Expand Down
20 changes: 10 additions & 10 deletions datafusion/core/src/datasource/listing/table.rs
Original file line number Diff line number Diff line change
Expand Up @@ -215,16 +215,16 @@ impl ListingTableConfig {
) -> Result<(String, Option<String>)> {
let mut exts = path.rsplit('.');

let splitted = exts.next().unwrap_or("");
let split = exts.next().unwrap_or("");

let file_compression_type = FileCompressionType::from_str(splitted)
let file_compression_type = FileCompressionType::from_str(split)
.unwrap_or(FileCompressionType::UNCOMPRESSED);

if file_compression_type.is_compressed() {
let splitted2 = exts.next().unwrap_or("");
Ok((splitted2.to_string(), Some(splitted.to_string())))
let split2 = exts.next().unwrap_or("");
Ok((split2.to_string(), Some(split.to_string())))
} else {
Ok((splitted.to_string(), None))
Ok((split.to_string(), None))
}
}

Expand Down Expand Up @@ -502,7 +502,7 @@ impl ListingOptions {
///
/// Currently this sets `target_partitions` and `collect_stat`
/// but if more options are added in the future that need to be coordinated
/// they will be synchronized thorugh this method.
/// they will be synchronized through this method.
pub fn with_session_config_options(mut self, config: &SessionConfig) -> Self {
self = self.with_target_partitions(config.target_partitions());
self = self.with_collect_stat(config.collect_statistics());
Expand Down Expand Up @@ -1132,8 +1132,8 @@ impl ListingTable {
}

// Expressions can be used for parttion pruning if they can be evaluated using
// only the partiton columns and there are partition columns.
fn can_be_evaluted_for_partition_pruning(
// only the partition columns and there are partition columns.
fn can_be_evaluated_for_partition_pruning(
partition_column_names: &[&str],
expr: &Expr,
) -> bool {
Expand Down Expand Up @@ -1182,7 +1182,7 @@ impl TableProvider for ListingTable {
// pushdown it to TableScan, otherwise, `unhandled` pruning predicates will be generated
let (partition_filters, filters): (Vec<_>, Vec<_>) =
filters.iter().cloned().partition(|filter| {
can_be_evaluted_for_partition_pruning(&table_partition_col_names, filter)
can_be_evaluated_for_partition_pruning(&table_partition_col_names, filter)
});

// We should not limit the number of partitioned files to scan if there are filters and limit
Expand Down Expand Up @@ -1271,7 +1271,7 @@ impl TableProvider for ListingTable {
filters
.iter()
.map(|filter| {
if can_be_evaluted_for_partition_pruning(&partition_column_names, filter)
if can_be_evaluated_for_partition_pruning(&partition_column_names, filter)
{
// if filter can be handled by partition pruning, it is exact
return Ok(TableProviderFilterPushDown::Exact);
Expand Down
8 changes: 4 additions & 4 deletions datafusion/core/src/datasource/physical_plan/parquet.rs
Original file line number Diff line number Diff line change
Expand Up @@ -332,7 +332,7 @@ mod tests {
let metric = get_value(&metrics, "pushdown_rows_pruned");
assert_eq!(metric, 3, "Expected all rows to be pruned");

// If we excplicitly allow nulls the rest of the predicate should work
// If we explicitly allow nulls the rest of the predicate should work
let filter = col("c2").is_null().and(col("c1").eq(lit(1_i32)));
let rt = RoundTrip::new()
.with_table_schema(table_schema.clone())
Expand Down Expand Up @@ -390,7 +390,7 @@ mod tests {
let metric = get_value(&metrics, "pushdown_rows_pruned");
assert_eq!(metric, 3, "Expected all rows to be pruned");

// If we excplicitly allow nulls the rest of the predicate should work
// If we explicitly allow nulls the rest of the predicate should work
let filter = col("c2").is_null().and(col("c1").eq(lit(1_i32)));
let rt = RoundTrip::new()
.with_table_schema(table_schema.clone())
Expand Down Expand Up @@ -452,7 +452,7 @@ mod tests {
let metric = get_value(&metrics, "pushdown_rows_pruned");
assert_eq!(metric, 3, "Expected all rows to be pruned");

// If we excplicitly allow nulls the rest of the predicate should work
// If we explicitly allow nulls the rest of the predicate should work
let filter = col("c2").is_null().and(col("c1").eq(lit(1_i32)));
let rt = RoundTrip::new()
.with_table_schema(table_schema.clone())
Expand Down Expand Up @@ -514,7 +514,7 @@ mod tests {
let metric = get_value(&metrics, "pushdown_rows_pruned");
assert_eq!(metric, 3, "Expected all rows to be pruned");

// If we excplicitly allow nulls the rest of the predicate should work
// If we explicitly allow nulls the rest of the predicate should work
let filter = col("c2").is_null().and(col("c3").eq(lit(7_i32)));
let rt = RoundTrip::new()
.with_table_schema(table_schema.clone())
Expand Down
2 changes: 1 addition & 1 deletion datafusion/core/src/physical_planner.rs
Original file line number Diff line number Diff line change
Expand Up @@ -2355,7 +2355,7 @@ impl<'n> TreeNodeVisitor<'n> for OptimizationInvariantChecker<'_> {

fn f_down(&mut self, node: &'n Self::Node) -> Result<TreeNodeRecursion> {
// Checks for the more permissive `InvariantLevel::Always`.
// Plans are not guarenteed to be executable after each physical optimizer run.
// Plans are not guaranteed to be executable after each physical optimizer run.
node.check_invariants(InvariantLevel::Always).map_err(|e|
e.context(format!("Invariant for ExecutionPlan node '{}' failed for PhysicalOptimizer rule '{}'", node.name(), self.rule.name()))
)?;
Expand Down
6 changes: 3 additions & 3 deletions datafusion/core/tests/dataframe/mod.rs
Original file line number Diff line number Diff line change
Expand Up @@ -92,8 +92,8 @@ async fn physical_plan_to_string(df: &DataFrame) -> String {
.await
.expect("Error creating physical plan");

let formated = displayable(physical_plan.as_ref()).indent(true);
formated.to_string()
let formatted = displayable(physical_plan.as_ref()).indent(true);
formatted.to_string()
}

pub fn table_with_constraints() -> Arc<dyn TableProvider> {
Expand Down Expand Up @@ -5660,7 +5660,7 @@ async fn test_alias() -> Result<()> {
.await?
.select(vec![col("a"), col("test.b"), lit(1).alias("one")])?
.alias("table_alias")?;
// All ouput column qualifiers are changed to "table_alias"
// All output column qualifiers are changed to "table_alias"
df.schema().columns().iter().for_each(|c| {
assert_eq!(c.relation, Some("table_alias".into()));
});
Expand Down
12 changes: 6 additions & 6 deletions datafusion/core/tests/execution/coop.rs
Original file line number Diff line number Diff line change
Expand Up @@ -434,7 +434,7 @@ async fn interleave_then_filter_all_yields(
let mut infinite_children = vec![];

// Use 32 distinct thresholds (each >0 and <8 192) to force 32 infinite inputs
for thr in 1..32 {
for threshold in 1..32 {
// One infinite exec:
let mut inf = make_lazy_exec_with_range("value", 0..i64::MAX, pretend_infinite);

Expand All @@ -444,7 +444,7 @@ async fn interleave_then_filter_all_yields(
let partitioning = Partitioning::Hash(exprs, 1);
inf.try_set_partitioning(partitioning)?;

// Apply a FilterExec: “(value / 8192) % thr == 0”.
// Apply a FilterExec: “(value / 8192) % threshold == 0”.
let filter_expr = binary(
binary(
binary(
Expand All @@ -454,7 +454,7 @@ async fn interleave_then_filter_all_yields(
&inf.schema(),
)?,
Modulo,
lit(thr as i64),
lit(threshold as i64),
&inf.schema(),
)?,
Eq,
Expand Down Expand Up @@ -490,7 +490,7 @@ async fn interleave_then_aggregate_yields(
let mut infinite_children = vec![];

// Use 32 distinct thresholds (each >0 and <8 192) to force 32 infinite inputs
for thr in 1..32 {
for threshold in 1..32 {
// One infinite exec:
let mut inf = make_lazy_exec_with_range("value", 0..i64::MAX, pretend_infinite);

Expand All @@ -500,7 +500,7 @@ async fn interleave_then_aggregate_yields(
let partitioning = Partitioning::Hash(exprs, 1);
inf.try_set_partitioning(partitioning)?;

// Apply a FilterExec: “(value / 8192) % thr == 0”.
// Apply a FilterExec: “(value / 8192) % threshold == 0”.
let filter_expr = binary(
binary(
binary(
Expand All @@ -510,7 +510,7 @@ async fn interleave_then_aggregate_yields(
&inf.schema(),
)?,
Modulo,
lit(thr as i64),
lit(threshold as i64),
&inf.schema(),
)?,
Eq,
Expand Down
2 changes: 1 addition & 1 deletion datafusion/core/tests/fuzz_cases/aggregate_fuzz.rs
Original file line number Diff line number Diff line change
Expand Up @@ -403,7 +403,7 @@ async fn run_aggregate_test(input1: Vec<RecordBatch>, group_by_columns: Vec<&str
Left Plan:\n{}\n\
Right Plan:\n{}\n\
schema:\n{schema}\n\
Left Ouptut:\n{}\n\
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

❤️

Left Output:\n{}\n\
Right Output:\n{}\n\
input:\n{}\n\
",
Expand Down
2 changes: 1 addition & 1 deletion datafusion/core/tests/fuzz_cases/pruning.rs
Original file line number Diff line number Diff line change
Expand Up @@ -201,7 +201,7 @@ impl Utf8Test {
}
}

/// all combinations of interesting charactes with lengths ranging from 1 to 4
/// all combinations of interesting characters with lengths ranging from 1 to 4
fn values() -> &'static [String] {
&VALUES
}
Expand Down
2 changes: 1 addition & 1 deletion datafusion/core/tests/fuzz_cases/sort_query_fuzz.rs
Original file line number Diff line number Diff line change
Expand Up @@ -220,7 +220,7 @@ impl SortQueryFuzzer {
.test_gen
.fuzzer_run(init_seed, query_seed, config_seed)
.await?;
println!("\n"); // Seperator between tested runs
println!("\n"); // Separator between tested runs

if expected_results.is_none() {
expected_results = Some(results);
Expand Down
10 changes: 5 additions & 5 deletions datafusion/core/tests/fuzz_cases/topk_filter_pushdown.rs
Original file line number Diff line number Diff line change
Expand Up @@ -219,16 +219,16 @@ struct RunQueryResult {
}

impl RunQueryResult {
fn expected_formated(&self) -> String {
fn expected_formatted(&self) -> String {
format!("{}", pretty_format_batches(&self.expected).unwrap())
}

fn result_formated(&self) -> String {
fn result_formatted(&self) -> String {
format!("{}", pretty_format_batches(&self.result).unwrap())
}

fn is_ok(&self) -> bool {
self.expected_formated() == self.result_formated()
self.expected_formatted() == self.result_formatted()
}
}

Expand Down Expand Up @@ -374,8 +374,8 @@ async fn test_fuzz_topk_filter_pushdown() {
for failure in &failures {
println!("Failure:");
println!("Query:\n{}", failure.query);
println!("\nExpected:\n{}", failure.expected_formated());
println!("\nResult:\n{}", failure.result_formated());
println!("\nExpected:\n{}", failure.expected_formatted());
println!("\nResult:\n{}", failure.result_formatted());
println!("\n\n");
}

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@ static INIT: Once = Once::new();

// ===========================================================================
// Test runners:
// Runners are splitted into multiple tests to run in parallel
// Runners are split into multiple tests to run in parallel
// ===========================================================================

#[test]
Expand Down
2 changes: 1 addition & 1 deletion datafusion/core/tests/memory_limit/mod.rs
Original file line number Diff line number Diff line change
Expand Up @@ -546,7 +546,7 @@ async fn test_external_sort_zero_merge_reservation() {
// Tests for disk limit (`max_temp_directory_size` in `DiskManager`)
// ------------------------------------------------------------------

// Create a new `SessionContext` with speicified disk limit, memory pool limit, and spill compression codec
// Create a new `SessionContext` with specified disk limit, memory pool limit, and spill compression codec
async fn setup_context(
disk_limit: u64,
memory_pool_limit: usize,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -441,7 +441,7 @@ impl TestConfig {

/// Perform a series of runs using the current [`TestConfig`],
/// assert the expected plan result,
/// and return the result plan (for potentional subsequent runs).
/// and return the result plan (for potential subsequent runs).
fn run(
&self,
expected_lines: &[&str],
Expand Down Expand Up @@ -2610,7 +2610,7 @@ fn parallelization_two_partitions_into_four() -> Result<()> {
"AggregateExec: mode=FinalPartitioned, gby=[a@0 as a], aggr=[]",
" RepartitionExec: partitioning=Hash([a@0], 4), input_partitions=4",
" AggregateExec: mode=Partial, gby=[a@0 as a], aggr=[]",
// Multiple source files splitted across partitions
// Multiple source files split across partitions
" DataSourceExec: file_groups={4 groups: [[x:0..50], [x:50..100], [y:0..50], [y:50..100]]}, projection=[a, b, c, d, e], file_type=parquet",
];
test_config.run(
Expand All @@ -2625,7 +2625,7 @@ fn parallelization_two_partitions_into_four() -> Result<()> {
"AggregateExec: mode=FinalPartitioned, gby=[a@0 as a], aggr=[]",
" RepartitionExec: partitioning=Hash([a@0], 4), input_partitions=4",
" AggregateExec: mode=Partial, gby=[a@0 as a], aggr=[]",
// Multiple source files splitted across partitions
// Multiple source files split across partitions
" DataSourceExec: file_groups={4 groups: [[x:0..50], [x:50..100], [y:0..50], [y:50..100]]}, projection=[a, b, c, d, e], file_type=csv, has_header=false",
];
test_config.run(&expected_csv, plan_csv.clone(), &DISTRIB_DISTRIB_SORT)?;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -236,7 +236,7 @@ async fn test_dynamic_filter_pushdown_through_hash_join_with_topk() {
config.optimizer.enable_dynamic_filter_pushdown = true;
config.execution.parquet.pushdown_filters = true;

// Appy the FilterPushdown optimizer rule
// Apply the FilterPushdown optimizer rule
let plan = FilterPushdown::new_post_optimization()
.optimize(Arc::clone(&plan), &config)
.unwrap();
Expand Down Expand Up @@ -603,7 +603,7 @@ fn test_no_pushdown_through_aggregates() {
}

/// Test various combinations of handling of child pushdown results
/// in an ExectionPlan in combination with support/not support in a DataSource.
/// in an ExecutionPlan in combination with support/not support in a DataSource.
#[test]
fn test_node_handles_child_pushdown_result() {
// If we set `with_support(true)` + `inject_filter = true` then the filter is pushed down to the DataSource
Expand Down
Loading