Skip to content

Commit 9d8cdb5

Browse files
committed
docs: remove some documentation from PyDataFrame
1 parent f0837de commit 9d8cdb5

File tree

1 file changed

+0
-51
lines changed

1 file changed

+0
-51
lines changed

src/dataframe.rs

Lines changed: 0 additions & 51 deletions
Original file line numberDiff line numberDiff line change
@@ -74,57 +74,6 @@ impl PyTableProvider {
7474
/// A PyDataFrame is a representation of a logical plan and an API to compose statements.
7575
/// Use it to build a plan and `.collect()` to execute the plan and collect the result.
7676
/// The actual execution of a plan runs natively on Rust and Arrow on a multi-threaded environment.
77-
///
78-
/// # Methods
79-
///
80-
/// - `new`: Creates a new PyDataFrame.
81-
/// - `__getitem__`: Enable selection for `df[col]`, `df[col1, col2, col3]`, and `df[[col1, col2, col3]]`.
82-
/// - `__repr__`: Returns a string representation of the DataFrame.
83-
/// - `_repr_html_`: Returns an HTML representation of the DataFrame.
84-
/// - `describe`: Calculate summary statistics for a DataFrame.
85-
/// - `schema`: Returns the schema from the logical plan.
86-
/// - `into_view`: Convert this DataFrame into a Table that can be used in register_table. We have not finalized on PyTableProvider approach yet.
87-
/// - `select_columns`: Select columns from the DataFrame.
88-
/// - `select`: Select expressions from the DataFrame.
89-
/// - `drop`: Drop columns from the DataFrame.
90-
/// - `filter`: Filter the DataFrame based on a predicate.
91-
/// - `with_column`: Add a new column to the DataFrame.
92-
/// - `with_columns`: Add multiple new columns to the DataFrame.
93-
/// - `with_column_renamed`: Rename a column in the DataFrame.
94-
/// - `aggregate`: Aggregate the DataFrame based on group by and aggregation expressions.
95-
/// - `sort`: Sort the DataFrame based on expressions.
96-
/// - `limit`: Limit the number of rows in the DataFrame.
97-
/// - `collect`: Executes the plan, returning a list of `RecordBatch`es.
98-
/// - `cache`: Cache the DataFrame.
99-
/// - `collect_partitioned`: Executes the DataFrame and collects all results into a vector of vector of RecordBatch maintaining the input partitioning.
100-
/// - `show`: Print the result, 20 lines by default.
101-
/// - `distinct`: Filter out duplicate rows.
102-
/// - `join`: Join two DataFrames.
103-
/// - `join_on`: Join two DataFrames based on expressions.
104-
/// - `explain`: Print the query plan.
105-
/// - `logical_plan`: Get the logical plan for this DataFrame.
106-
/// - `optimized_logical_plan`: Get the optimized logical plan for this DataFrame.
107-
/// - `execution_plan`: Get the execution plan for this DataFrame.
108-
/// - `repartition`: Repartition the DataFrame based on a logical partitioning scheme.
109-
/// - `repartition_by_hash`: Repartition the DataFrame based on a hash partitioning scheme.
110-
/// - `union`: Calculate the union of two DataFrames, preserving duplicate rows.
111-
/// - `union_distinct`: Calculate the distinct union of two DataFrames.
112-
/// - `unnest_column`: Unnest a column in the DataFrame.
113-
/// - `unnest_columns`: Unnest multiple columns in the DataFrame.
114-
/// - `intersect`: Calculate the intersection of two DataFrames.
115-
/// - `except_all`: Calculate the exception of two DataFrames.
116-
/// - `write_csv`: Write the DataFrame to a CSV file.
117-
/// - `write_parquet`: Write the DataFrame to a Parquet file.
118-
/// - `write_json`: Write the DataFrame to a JSON file.
119-
/// - `to_arrow_table`: Convert the DataFrame to an Arrow Table.
120-
/// - `__arrow_c_stream__`: Convert the DataFrame to an Arrow C Stream.
121-
/// - `execute_stream`: Execute the DataFrame and return a RecordBatchStream.
122-
/// - `execute_stream_partitioned`: Execute the DataFrame and return partitioned RecordBatchStreams.
123-
/// - `to_pandas`: Convert the DataFrame to a Pandas DataFrame.
124-
/// - `to_pylist`: Convert the DataFrame to a Python list.
125-
/// - `to_pydict`: Convert the DataFrame to a Python dictionary.
126-
/// - `to_polars`: Convert the DataFrame to a Polars DataFrame.
127-
/// - `count`: Execute the DataFrame to get the total number of rows.
12877
#[pyclass(name = "DataFrame", module = "datafusion", subclass)]
12978
#[derive(Clone)]
13079
pub struct PyDataFrame {

0 commit comments

Comments
 (0)