You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/interactive-etl/index.md
+5-5Lines changed: 5 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -50,7 +50,7 @@ In order to run this example you will need:
50
50
1. Deploy a [Flink session cluster](https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#session-cluster-deployments) (used for long-running, multipurpose deployments) using the [Flink Kubernetes Operator](https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/):
This might take a little while to show, there aren't many people buying lots of units.
162
-
Ok, lets find the true big spenders.
163
-
User IDs forusers who buy 3 or more items over 500 GBP:
162
+
Next, lets find the popular products which are also expensive.
163
+
Product IDs forproducts that cost more than £500 and which someone bought at least 3 units of at once:
164
164
165
165
```sql
166
166
SELECT
@@ -336,11 +336,11 @@ The ETL query (deployed above) will run like any other Flink streaming job and c
336
336
However, your session Flink cluster might primarily be for data exploration and development, which means your ETL job would be competing for resources with other queries.
337
337
If your transformed data is needed in production, it would be better to deploy the query as a stand-alone Flink Job independent of the session Flink cluster.
338
338
339
-
There is an example FlinkDeployment CR (`standalone-etl-deployment.yaml`) in this directory that will deploy the queries above in Flink's application mode.
339
+
There is an example FlinkDeployment CR (`standalone-etl-deployment.yaml`) that will deploy the queries above in Flink's application mode.
340
340
This will deploy the ETL query in a self-contained Flink cluster that can be managed like any other FlinkDeployment.
Once you know that is running (`kubectl -n flink get pods`), you can see the cleaned data in Kafka by querying the new output topic (this has a different name to the one used in the interactive demo):
0 commit comments