You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/BigQueryPushdownEngine-sqlengine.md
+15-11Lines changed: 15 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -73,17 +73,6 @@ completed. This API can be used if the execution environment for this environmen
73
73
Note that this API has an on-demand price model. See the [Pricing](https://cloud.google.com/bigquery/pricing#storage-api)
74
74
page for details related to pricing.
75
75
76
-
**Attempt direct copy to BigQuery sinks**: Performance can be greatly improved if the records from stages that are executed using
77
-
BigQuery ELT Transformation Pushdown are copied directly into a configured table in a compatible BigQuery Sink. This eliminates
78
-
the need to read records into the pipeline as no further processing is needed within the pipeline.
79
-
To ensure this BigQuery Sink can take advantage of the performance improvements provided by this feature,
80
-
the following requirements must be met:
81
-
1. The service account configured for BigQuery ELT Transformation Pushdown has permissions to create and update tables in the dataset used by the BigQuery Sink.
82
-
2. The datasets used for BigQuery ELT Transformation Pushdown and BigQuery Sink must be stored in the same **location**.
83
-
3. The **operation** is either `insert` (With **Truncate Table** disabled), `update`or `upsert`
84
-
85
-
Note: If the direct copy operation does not succeed, the pipeline will proceed with the standard workflow in order to ensure completion.
86
-
87
76
**Service Account** - service account key used for authorization
88
77
89
78
***File Path**: Path on the local file system of the service account key used for
@@ -121,6 +110,21 @@ corresponding BigQuery data type for each CDAP type, for updates and upserts.
121
110
If any of the stages involved in a Join operation contains an unsupported type,
122
111
this Join operation will be executed in Spark.
123
112
113
+
Writing to BigQuery Sinks
114
+
-------------------------
115
+
116
+
Performance can be greatly improved if the records from stages that are executed using BigQuery Transformation
117
+
Pushdown are copied directly into a configured table in a compatible BigQuery Sink.
118
+
This eliminates the need to read records into the pipeline as no further processing is needed within the pipeline.
119
+
120
+
To ensure this BigQuery Sink can take advantage of the performance improvements provided by this feature,
121
+
the following requirements must be met:
122
+
1. The service account configured for BigQuery ELT Transformation Pushdown has permissions to create and update tables in the dataset used by the BigQuery Sink.
123
+
2. The datasets used for BigQuery ELT Transformation Pushdown and BigQuery Sink must be stored in the same **location**.
124
+
3. The **operation** is either `insert` (With **Truncate Table** disabled), `update`or `upsert`
125
+
126
+
Note: If the direct copy operation does not succeed, the pipeline will proceed with the standard workflow in order to ensure completion.
0 commit comments