Skip to content

Commit 9b16f40

Browse files
committed
Merge branch 'develop'
2 parents 3b6bc23 + ee8cd30 commit 9b16f40

File tree

124 files changed

+2961
-1201
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

124 files changed

+2961
-1201
lines changed

.github/workflows/get-matrix.yml

Lines changed: 23 additions & 23 deletions
Original file line numberDiff line numberDiff line change
@@ -86,7 +86,7 @@ jobs:
8686

8787
- name: Check if base files are changed
8888
id: changed-base
89-
uses: tj-actions/changed-files@v45
89+
uses: tj-actions/changed-files@v46
9090
with:
9191
files_from_source_file: .github/workflows/data/base/tracked.txt
9292
files_ignore_from_source_file: .github/workflows/data/base/ignored.txt
@@ -97,7 +97,7 @@ jobs:
9797
9898
- name: Check if db-related files are changed
9999
id: changed-db
100-
uses: tj-actions/changed-files@v45
100+
uses: tj-actions/changed-files@v46
101101
with:
102102
files_from_source_file: .github/workflows/data/db/tracked.txt
103103
files_ignore_from_source_file: .github/workflows/data/db/ignored.txt
@@ -108,7 +108,7 @@ jobs:
108108
109109
- name: Check if file-related files are changed
110110
id: changed-file
111-
uses: tj-actions/changed-files@v45
111+
uses: tj-actions/changed-files@v46
112112
with:
113113
files_from_source_file: .github/workflows/data/file/tracked.txt
114114
files_ignore_from_source_file: .github/workflows/data/file/ignored.txt
@@ -119,7 +119,7 @@ jobs:
119119
120120
- name: Check if file-df-related files are changed
121121
id: changed-file-df
122-
uses: tj-actions/changed-files@v45
122+
uses: tj-actions/changed-files@v46
123123
with:
124124
files_from_source_file: .github/workflows/data/file-df/tracked.txt
125125
files_ignore_from_source_file: .github/workflows/data/file-df/ignored.txt
@@ -130,7 +130,7 @@ jobs:
130130
131131
- name: Check if core files are changed
132132
id: changed-core
133-
uses: tj-actions/changed-files@v45
133+
uses: tj-actions/changed-files@v46
134134
with:
135135
files_from_source_file: .github/workflows/data/core/tracked.txt
136136
files_ignore_from_source_file: .github/workflows/data/core/ignored.txt
@@ -160,7 +160,7 @@ jobs:
160160

161161
- name: Check if Clickhouse files are changed
162162
id: changed-clickhouse
163-
uses: tj-actions/changed-files@v45
163+
uses: tj-actions/changed-files@v46
164164
with:
165165
files_from_source_file: .github/workflows/data/clickhouse/tracked.txt
166166
files_ignore_from_source_file: .github/workflows/data/clickhouse/ignored.txt
@@ -190,7 +190,7 @@ jobs:
190190

191191
- name: Check if Greenplum files are changed
192192
id: changed-greenplum
193-
uses: tj-actions/changed-files@v45
193+
uses: tj-actions/changed-files@v46
194194
with:
195195
files_from_source_file: .github/workflows/data/greenplum/tracked.txt
196196
files_ignore_from_source_file: .github/workflows/data/greenplum/ignored.txt
@@ -220,7 +220,7 @@ jobs:
220220

221221
- name: Check if Hive files are changed
222222
id: changed-hive
223-
uses: tj-actions/changed-files@v45
223+
uses: tj-actions/changed-files@v46
224224
with:
225225
files_from_source_file: .github/workflows/data/hive/tracked.txt
226226
files_ignore_from_source_file: .github/workflows/data/hive/ignored.txt
@@ -250,7 +250,7 @@ jobs:
250250

251251
- name: Check if Kafka files are changed
252252
id: changed-kafka
253-
uses: tj-actions/changed-files@v45
253+
uses: tj-actions/changed-files@v46
254254
with:
255255
files_from_source_file: .github/workflows/data/kafka/tracked.txt
256256
files_ignore_from_source_file: .github/workflows/data/kafka/ignored.txt
@@ -280,7 +280,7 @@ jobs:
280280

281281
- name: Check if LocalFS files are changed
282282
id: changed-local-fs
283-
uses: tj-actions/changed-files@v45
283+
uses: tj-actions/changed-files@v46
284284
with:
285285
files_from_source_file: .github/workflows/data/local-fs/tracked.txt
286286
files_ignore_from_source_file: .github/workflows/data/local-fs/ignored.txt
@@ -310,7 +310,7 @@ jobs:
310310

311311
- name: Check if MongoDB files are changed
312312
id: changed-mongodb
313-
uses: tj-actions/changed-files@v45
313+
uses: tj-actions/changed-files@v46
314314
with:
315315
files_from_source_file: .github/workflows/data/mongodb/tracked.txt
316316
files_ignore_from_source_file: .github/workflows/data/mongodb/ignored.txt
@@ -340,7 +340,7 @@ jobs:
340340

341341
- name: Check if MSSQL files are changed
342342
id: changed-mssql
343-
uses: tj-actions/changed-files@v45
343+
uses: tj-actions/changed-files@v46
344344
with:
345345
files_from_source_file: .github/workflows/data/mssql/tracked.txt
346346
files_ignore_from_source_file: .github/workflows/data/mssql/ignored.txt
@@ -370,7 +370,7 @@ jobs:
370370

371371
- name: Check if MySQL files are changed
372372
id: changed-mysql
373-
uses: tj-actions/changed-files@v45
373+
uses: tj-actions/changed-files@v46
374374
with:
375375
files_from_source_file: .github/workflows/data/mysql/tracked.txt
376376
files_ignore_from_source_file: .github/workflows/data/mysql/ignored.txt
@@ -400,7 +400,7 @@ jobs:
400400

401401
- name: Check if Oracle files are changed
402402
id: changed-oracle
403-
uses: tj-actions/changed-files@v45
403+
uses: tj-actions/changed-files@v46
404404
with:
405405
files_from_source_file: .github/workflows/data/oracle/tracked.txt
406406
files_ignore_from_source_file: .github/workflows/data/oracle/ignored.txt
@@ -430,7 +430,7 @@ jobs:
430430

431431
- name: Check if Postgres files are changed
432432
id: changed-postgres
433-
uses: tj-actions/changed-files@v45
433+
uses: tj-actions/changed-files@v46
434434
with:
435435
files_from_source_file: .github/workflows/data/postgres/tracked.txt
436436
files_ignore_from_source_file: .github/workflows/data/postgres/ignored.txt
@@ -460,7 +460,7 @@ jobs:
460460

461461
- name: Check if Teradata files are changed
462462
id: changed-teradata
463-
uses: tj-actions/changed-files@v45
463+
uses: tj-actions/changed-files@v46
464464
with:
465465
files_from_source_file: .github/workflows/data/teradata/tracked.txt
466466
files_ignore_from_source_file: .github/workflows/data/teradata/ignored.txt
@@ -490,7 +490,7 @@ jobs:
490490

491491
- name: Check if FTP files are changed
492492
id: changed-ftp
493-
uses: tj-actions/changed-files@v45
493+
uses: tj-actions/changed-files@v46
494494
with:
495495
files_from_source_file: .github/workflows/data/ftp/tracked.txt
496496
files_ignore_from_source_file: .github/workflows/data/ftp/ignored.txt
@@ -520,7 +520,7 @@ jobs:
520520

521521
- name: Check if FTPS files are changed
522522
id: changed-ftps
523-
uses: tj-actions/changed-files@v45
523+
uses: tj-actions/changed-files@v46
524524
with:
525525
files_from_source_file: .github/workflows/data/ftps/tracked.txt
526526
files_ignore_from_source_file: .github/workflows/data/ftps/ignored.txt
@@ -550,7 +550,7 @@ jobs:
550550

551551
- name: Check if HDFS files are changed
552552
id: changed-hdfs
553-
uses: tj-actions/changed-files@v45
553+
uses: tj-actions/changed-files@v46
554554
with:
555555
files_from_source_file: .github/workflows/data/hdfs/tracked.txt
556556
files_ignore_from_source_file: .github/workflows/data/hdfs/ignored.txt
@@ -580,7 +580,7 @@ jobs:
580580

581581
- name: Check if S3 files are changed
582582
id: changed-s3
583-
uses: tj-actions/changed-files@v45
583+
uses: tj-actions/changed-files@v46
584584
with:
585585
files_from_source_file: .github/workflows/data/s3/tracked.txt
586586
files_ignore_from_source_file: .github/workflows/data/s3/ignored.txt
@@ -610,7 +610,7 @@ jobs:
610610

611611
- name: Check if SFTP files are changed
612612
id: changed-sftp
613-
uses: tj-actions/changed-files@v45
613+
uses: tj-actions/changed-files@v46
614614
with:
615615
files_from_source_file: .github/workflows/data/sftp/tracked.txt
616616
files_ignore_from_source_file: .github/workflows/data/sftp/ignored.txt
@@ -640,7 +640,7 @@ jobs:
640640

641641
- name: Check if Samba files are changed
642642
id: changed-samba
643-
uses: tj-actions/changed-files@v45
643+
uses: tj-actions/changed-files@v46
644644
with:
645645
files_from_source_file: .github/workflows/data/samba/tracked.txt
646646
files_ignore_from_source_file: .github/workflows/data/samba/ignored.txt
@@ -670,7 +670,7 @@ jobs:
670670

671671
- name: Check if WebDAV files are changed
672672
id: changed-webdav
673-
uses: tj-actions/changed-files@v45
673+
uses: tj-actions/changed-files@v46
674674
with:
675675
files_from_source_file: .github/workflows/data/webdav/tracked.txt
676676
files_ignore_from_source_file: .github/workflows/data/webdav/ignored.txt

docs/changelog/0.13.4.rst

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,14 @@
1+
0.13.4 (2025-03-20)
2+
===================
3+
4+
Doc only Changes
5+
----------------
6+
7+
- Prefer ``ReadOptions(partitionColumn=..., numPartitions=..., queryTimeout=...)``
8+
instead of ``ReadOptions(partition_column=..., num_partitions=..., query_timeout=...)``,
9+
to match Spark documentation. (:github:pull:`352`)
10+
11+
- Prefer ``WriteOptions(if_exists=...)`` instead of ``WriteOptions(mode=...)`` for IDE suggestions. (:github:pull:`354`)
12+
13+
- Document all options of supported file formats.
14+
(:github:pull:`355`, :github:pull:`356`, :github:pull:`357`, :github:pull:`358`, :github:pull:`359`, :github:pull:`360`, :github:pull:`361`, :github:pull:`362`)

docs/changelog/index.rst

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -3,6 +3,7 @@
33
:caption: Changelog
44

55
DRAFT
6+
0.13.4
67
0.13.3
78
0.13.1
89
0.13.0

docs/connection/db_connection/clickhouse/execute.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ Examples
5050
5151
df = clickhouse.fetch(
5252
"SELECT value FROM some.reference_table WHERE key = 'some_constant'",
53-
options=Clickhouse.FetchOptions(query_timeout=10),
53+
options=Clickhouse.FetchOptions(queryTimeout=10),
5454
)
5555
clickhouse.close()
5656
value = df.collect()[0][0] # get value from first row and first column
@@ -96,7 +96,7 @@ Examples
9696
ENGINE = MergeTree()
9797
ORDER BY id
9898
""",
99-
options=Clickhouse.ExecuteOptions(query_timeout=10),
99+
options=Clickhouse.ExecuteOptions(queryTimeout=10),
100100
)
101101
102102
Notes

docs/connection/db_connection/clickhouse/read.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@ Snapshot strategy:
4141
source="schema.table",
4242
columns=["id", "key", "CAST(value AS String) value", "updated_dt"],
4343
where="key = 'something'",
44-
options=Clickhouse.ReadOptions(partition_column="id", num_partitions=10),
44+
options=Clickhouse.ReadOptions(partitionColumn="id", numPartitions=10),
4545
)
4646
df = reader.run()
4747
@@ -61,7 +61,7 @@ Incremental strategy:
6161
columns=["id", "key", "CAST(value AS String) value", "updated_dt"],
6262
where="key = 'something'",
6363
hwm=DBReader.AutoDetectHWM(name="clickhouse_hwm", expression="updated_dt"),
64-
options=Clickhouse.ReadOptions(partition_column="id", num_partitions=10),
64+
options=Clickhouse.ReadOptions(partitionColumn="id", numPartitions=10),
6565
)
6666
6767
with IncrementalStrategy():

docs/connection/db_connection/clickhouse/sql.rst

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -44,10 +44,10 @@ Examples
4444
key = 'something'
4545
""",
4646
options=Clickhouse.SQLOptions(
47-
partition_column="id",
48-
num_partitions=10,
49-
lower_bound=0,
50-
upper_bound=1000,
47+
partitionColumn="id",
48+
numPartitions=10,
49+
lowerBound=0,
50+
upperBound=1000,
5151
),
5252
)
5353

docs/connection/db_connection/clickhouse/types.rst

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -77,7 +77,7 @@ So instead of relying on Spark to create tables:
7777
7878
writer = DBWriter(
7979
connection=clickhouse,
80-
table="default.target_tbl",
80+
target="default.target_tbl",
8181
options=Clickhouse.WriteOptions(
8282
if_exists="append",
8383
# ENGINE is required by Clickhouse
@@ -105,7 +105,7 @@ Always prefer creating tables with specific types **BEFORE WRITING DATA**:
105105
106106
writer = DBWriter(
107107
connection=clickhouse,
108-
table="default.target_tbl",
108+
target="default.target_tbl",
109109
options=Clickhouse.WriteOptions(if_exists="append"),
110110
)
111111
writer.run(df)
@@ -375,7 +375,7 @@ For parsing JSON columns in ClickHouse, :obj:`JSON.parse_column <onetl.file.form
375375
376376
reader = DBReader(
377377
connection=clickhouse,
378-
table="default.source_tbl",
378+
target="default.source_tbl",
379379
columns=[
380380
"id",
381381
"toJSONString(array_column) array_column",

docs/connection/db_connection/greenplum/execute.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ Examples
5050
5151
df = greenplum.fetch(
5252
"SELECT value FROM some.reference_table WHERE key = 'some_constant'",
53-
options=Greenplum.FetchOptions(query_timeout=10),
53+
options=Greenplum.FetchOptions(queryTimeout=10),
5454
)
5555
greenplum.close()
5656
value = df.collect()[0][0] # get value from first row and first column
@@ -97,7 +97,7 @@ Examples
9797
)
9898
DISTRIBUTED BY id
9999
""",
100-
options=Greenplum.ExecuteOptions(query_timeout=10),
100+
options=Greenplum.ExecuteOptions(queryTimeout=10),
101101
)
102102
103103
Interaction schema

docs/connection/db_connection/greenplum/read.rst

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -275,11 +275,11 @@ In this case, custom column can be used instead:
275275
reader = DBReader(
276276
connection=greenplum,
277277
source="schema.view_with_partition_column",
278-
options=Greenplum.Options(
278+
options=Greenplum.ReadOptions(
279279
# parallelize data using specified column
280-
partition_column="part_column",
280+
partitionColumn="part_column",
281281
# create 10 Spark tasks, each will read only part of table data
282-
num_partitions=10,
282+
partitions=10,
283283
),
284284
)
285285
df = reader.run()
@@ -348,8 +348,8 @@ and then read this table using ``DBReader``:
348348

349349
.. code-block:: python
350350
351-
df1 = DBReader(connection=greenplum, table="public.table1", ...).run()
352-
df2 = DBReader(connection=greenplum, table="public.table2", ...).run()
351+
df1 = DBReader(connection=greenplum, target="public.table1", ...).run()
352+
df2 = DBReader(connection=greenplum, target="public.table2", ...).run()
353353
354354
joined_df = df1.join(df2, on="col")
355355

docs/connection/db_connection/greenplum/types.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -66,7 +66,7 @@ So instead of relying on Spark to create tables:
6666
6767
writer = DBWriter(
6868
connection=greenplum,
69-
table="public.table",
69+
target="public.table",
7070
options=Greenplum.WriteOptions(
7171
if_exists="append",
7272
# by default distribution is random
@@ -96,7 +96,7 @@ Always prefer creating table with desired DDL **BEFORE WRITING DATA**:
9696
9797
writer = DBWriter(
9898
connection=greenplum,
99-
table="public.table",
99+
target="public.table",
100100
options=Greenplum.WriteOptions(if_exists="append"),
101101
)
102102
writer.run(df)

0 commit comments

Comments
 (0)