Skip to content

Commit 2f5212f

Browse files
committed
chore(release): Update image references with stackable24.11.1
1 parent 1757cda commit 2f5212f

32 files changed

+42
-42
lines changed

demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: start-pyspark-job
11-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0
11+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1
1212
# N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly
1313
# restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps
1414
# below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that

demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: start-date-job
11-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0
11+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1
1212
# N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly
1313
# restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps
1414
# below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that

demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,11 +9,11 @@ spec:
99
serviceAccountName: demo-serviceaccount
1010
initContainers:
1111
- name: wait-for-kafka
12-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0
12+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1
1313
command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/instance=kafka -l app.kubernetes.io/name=kafka"]
1414
containers:
1515
- name: create-nifi-ingestion-job
16-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0
16+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1
1717
command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/LakehouseKafkaIngest.xml && python -u /tmp/script/script.py"]
1818
volumeMounts:
1919
- name: script

demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,11 +12,11 @@ spec:
1212
serviceAccountName: demo-serviceaccount
1313
initContainers:
1414
- name: wait-for-kafka
15-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0
15+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1
1616
command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/name=kafka -l app.kubernetes.io/instance=kafka"]
1717
containers:
1818
- name: create-spark-ingestion-job
19-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0
19+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1
2020
command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ingestion-job.yaml"]
2121
volumeMounts:
2222
- name: manifest

demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,11 +9,11 @@ spec:
99
serviceAccountName: demo-serviceaccount
1010
initContainers:
1111
- name: wait-for-testdata
12-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0
12+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1
1313
command: ["bash", "-c", "echo 'Waiting for job load-test-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-test-data"]
1414
containers:
1515
- name: create-tables-in-trino
16-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0
16+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1
1717
command: ["bash", "-c", "python -u /tmp/script/script.py"]
1818
volumeMounts:
1919
- name: script

demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: setup-superset
11-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0
11+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1
1212
command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/superset-assets.zip && python -u /tmp/script/script.py"]
1313
volumeMounts:
1414
- name: script

demos/end-to-end-security/create-spark-report.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ spec:
1212
serviceAccountName: demo-serviceaccount
1313
initContainers:
1414
- name: wait-for-trino-tables
15-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0
15+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1
1616
command:
1717
- bash
1818
- -euo
@@ -23,7 +23,7 @@ spec:
2323
kubectl wait --timeout=30m --for=condition=complete job/create-tables-in-trino
2424
containers:
2525
- name: create-spark-report
26-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0
26+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1
2727
command:
2828
- bash
2929
- -euo

demos/end-to-end-security/create-trino-tables.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: create-tables-in-trino
11-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0
11+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1
1212
command: ["bash", "-c", "python -u /tmp/script/script.py"]
1313
volumeMounts:
1414
- name: script

demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ spec:
99
spec:
1010
containers:
1111
- name: create-hfile-and-import-to-hbase
12-
image: docker.stackable.tech/stackable/hbase:2.4.18-stackable24.11.0
12+
image: docker.stackable.tech/stackable/hbase:2.4.18-stackable24.11.1
1313
env:
1414
- name: HADOOP_USER_NAME
1515
value: stackable

demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: load-ny-taxi-data
11-
image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable24.11.0
11+
image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable24.11.1
1212
# yamllint disable rule:line-length
1313
command: ["bash", "-c", "/stackable/hadoop/bin/hdfs dfs -mkdir -p /ny-taxi-data/raw \
1414
&& cd /tmp \

0 commit comments

Comments
 (0)