You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
description: Location root for Synapse Link data (Dataverse container)
81
+
description: Location root for Synapse Link data (Dataverse container), in a format abfss://account@container.dfs.core.windows.net/
82
82
changeCaptureIntervalSeconds:
83
83
type: integer
84
-
description: How long to wait before polling for next result set. Can be from 1 to 1 hour.
84
+
description: How long to wait before polling for next change set. Accepted range is between 1s and 3600s
85
85
minimum: 1
86
86
maximum: 3600
87
87
connectionStringRef:
@@ -93,7 +93,7 @@ spec:
93
93
type: string
94
94
jobTemplateRef:
95
95
description: |
96
-
Name of the job template to be used for the streaming job if stream is running in normal mode.
96
+
Name of the job template to be used for the application in streaming mode.
97
97
type: object
98
98
properties:
99
99
name:
@@ -104,7 +104,7 @@ spec:
104
104
type: string
105
105
backfillJobTemplateRef:
106
106
description: |
107
-
Name of the job template to be used for the streaming job if stream is running in the backfill mode.
107
+
Name of the job template to be used for the for the application in backfill mode.
108
108
type: object
109
109
properties:
110
110
name:
@@ -121,42 +121,42 @@ spec:
121
121
description: Max retry delay on blob reads for the http client.
122
122
rowsPerGroup:
123
123
type: integer
124
-
description: Number of rows per parquet rowgroup.
124
+
description: Maximum number of rows to be grouped together for the staging process to consume.
125
125
groupingIntervalSeconds:
126
126
type: integer
127
-
description: Max time to wait for rowsPerGroup to accumulate. Can be from 1 to 60 seconds.
127
+
description: Max time to wait for rowsPerGroup to accumulate and then proceed to staging. Can be from 1 to 60 seconds.
128
128
minimum: 1
129
129
maximum: 60
130
130
sinkSettings:
131
131
type: object
132
132
properties:
133
133
optimizeSettings:
134
134
type: object
135
-
description: Optimization settings for Iceberg tables.
135
+
description: Configuration for target table optimize (data file aggregation into bigger files)
136
136
properties:
137
137
batchThreshold:
138
138
type: integer
139
139
default: 60
140
-
description: Number of batches to accumulate before running the optimization query.
140
+
description: Number of batches to accumulate before triggering the OPTIMIZE
141
141
fileSizeThreshold:
142
142
type: string
143
143
default: 100MB
144
-
description: File size to target for the optimization query.
144
+
description: File size to target when running OPTIMIZE
145
145
default:
146
146
batchThreshold: 60
147
147
fileSizeThreshold: 100MB
148
148
snapshotExpirationSettings:
149
149
type: object
150
-
description: Expiration query configuration for Iceberg tables.
150
+
description: Configuration for EXPIRE SNAPSHOTS (table transaction log cutoff)
151
151
properties:
152
152
batchThreshold:
153
153
type: integer
154
154
default: 60
155
-
description: Number of batches to accumulate before running the snapshot expiration query.
155
+
description: Number of batches to accumulate triggering EXPIRE SNAPSHOTS
156
156
retentionThreshold:
157
157
type: string
158
158
default: 6h
159
-
description: File retention period.
159
+
description: Maximum age of records in the transaction log to keep
160
160
default:
161
161
batchThreshold: 60
162
162
retentionThreshold: 6h
@@ -167,17 +167,17 @@ spec:
167
167
batchThreshold:
168
168
type: integer
169
169
default: 60
170
-
description: Number of batches to accumulate before running the expire orphan files query.
170
+
description: Configuration for EXPIRE ORPHAN FILES (cleanup of files no longer referenced by Iceberg snapshots)
171
171
retentionThreshold:
172
172
type: string
173
173
default: 6h
174
-
description: File retention period.
174
+
description: Number of batches to accumulate before triggering EXPIRE ORPHAN FILES
175
175
default:
176
176
batchThreshold: 60
177
177
retentionThreshold: 6h
178
178
analyzeSettings:
179
179
type: object
180
-
description: Settings for running ANALYZE on target.
180
+
description: Configuration for ANALYZE (full refresh of extended statistics on the target)
181
181
properties:
182
182
batchThreshold:
183
183
type: integer
@@ -195,9 +195,21 @@ spec:
195
195
targetTableName:
196
196
type: string
197
197
description: Name for the target Iceberg table.
198
+
sinkCatalogSettings:
199
+
type: object
200
+
description: Connection settings for Iceberg REST Catalog for the sink (target). This is used by watermarking process.
201
+
properties:
202
+
namespace:
203
+
type: string
204
+
warehouse:
205
+
type: string
206
+
catalogUri:
207
+
type: string
198
208
lookBackInterval:
199
209
type: integer
200
210
description: |
211
+
DEPRECATED - TO BE REMOVED IN 2.1 RELEASE. DO NOT USE THIS SETTING!
212
+
201
213
Number of seconds to look back when determining first set of changes to extract.
202
214
Can be set in interval from 1 second to 10 hours. Default is 1 hour.
203
215
minimum: 1
@@ -206,15 +218,19 @@ spec:
206
218
stagingDataSettings:
207
219
type: object
208
220
properties:
221
+
dataLocation:
222
+
type: string
223
+
description: Option data location override. Only use this setting for debugging and never in production environments.
209
224
maxRowsPerFile:
210
225
type: integer
226
+
description: The maximum number of rows per each data file in the staging table.
211
227
default: 10000
212
-
dataLocation:
213
-
type: string
214
228
tableNamePrefix:
215
229
type: string
230
+
description: Prefix for staging tables created by Arcane. Must be UNIQUE in the WAREHOUSE scope.
216
231
catalog:
217
232
type: object
233
+
description: Settings for Iceberg REST Catalog used for staging tables
218
234
properties:
219
235
catalogName:
220
236
type: string
@@ -233,6 +249,7 @@ spec:
233
249
overwrite
234
250
fieldSelectionRule:
235
251
type: object
252
+
description: INCLUDE will only use fields provided. You must specify mandatory fields like ARCANE_MERGE_KEY as well. EXCLUDE will exclude provided fields instead. ALL (default) will use all fields without filters.
0 commit comments