Skip to content
Merged
Show file tree
Hide file tree
Changes from 2 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: start-pyspark-job
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0
image: oci.stackable.tech/sdp/tools:1.0.0-stackable24.7.0
# N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly
# restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps
# below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: start-date-job
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0
image: oci.stackable.tech/sdp/tools:1.0.0-stackable24.7.0
# N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly
# restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps
# below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -9,11 +9,11 @@ spec:
serviceAccountName: demo-serviceaccount
initContainers:
- name: wait-for-kafka
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0
image: oci.stackable.tech/sdp/tools:1.0.0-stackable24.7.0
command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/instance=kafka -l app.kubernetes.io/name=kafka"]
containers:
- name: create-nifi-ingestion-job
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/LakehouseKafkaIngest.xml && python -u /tmp/script/script.py"]
volumeMounts:
- name: script
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -12,11 +12,11 @@ spec:
serviceAccountName: demo-serviceaccount
initContainers:
- name: wait-for-kafka
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0
image: oci.stackable.tech/sdp/tools:1.0.0-stackable24.7.0
command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/name=kafka -l app.kubernetes.io/instance=kafka"]
containers:
- name: create-spark-ingestion-job
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0
image: oci.stackable.tech/sdp/tools:1.0.0-stackable24.7.0
command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ingestion-job.yaml"]
volumeMounts:
- name: manifest
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -9,11 +9,11 @@ spec:
serviceAccountName: demo-serviceaccount
initContainers:
- name: wait-for-testdata
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0
image: oci.stackable.tech/sdp/tools:1.0.0-stackable24.7.0
command: ["bash", "-c", "echo 'Waiting for job load-test-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-test-data"]
containers:
- name: create-tables-in-trino
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "python -u /tmp/script/script.py"]
volumeMounts:
- name: script
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: setup-superset
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/superset-assets.zip && python -u /tmp/script/script.py"]
volumeMounts:
- name: script
Expand Down
4 changes: 2 additions & 2 deletions demos/end-to-end-security/create-spark-report.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ spec:
serviceAccountName: demo-serviceaccount
initContainers:
- name: wait-for-trino-tables
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command:
- bash
- -euo
Expand All @@ -23,7 +23,7 @@ spec:
kubectl wait --timeout=30m --for=condition=complete job/create-tables-in-trino
containers:
- name: create-spark-report
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command:
- bash
- -euo
Expand Down
2 changes: 1 addition & 1 deletion demos/end-to-end-security/create-trino-tables.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: create-tables-in-trino
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "python -u /tmp/script/script.py"]
volumeMounts:
- name: script
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ spec:
spec:
containers:
- name: create-hfile-and-import-to-hbase
image: docker.stackable.tech/stackable/hbase:2.4.18-stackable24.7.0
image: oci.stackable.tech/sdp/hbase:2.4.18-stackable24.7.0
env:
- name: HADOOP_USER_NAME
value: stackable
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ spec:
# We use 24.3.0 here which contains the distcp MapReduce components
# This is not included in the 24.7 and 24.11 images and will fail.
# See: https://github.com/stackabletech/docker-images/issues/793
image: docker.stackable.tech/stackable/hadoop:3.3.6-stackable24.3.0
image: oci.stackable.tech/sdp/hadoop:3.3.6-stackable24.3.0
env:
- name: HADOOP_USER_NAME
value: stackable
Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM docker.stackable.tech/stackable/spark-k8s:3.5.0-stackable24.3.0
FROM oci.stackable.tech/sdp/spark-k8s:3.5.0-stackable24.3.0

COPY demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/requirements.txt .

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: load-ny-taxi-data
image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable0.0.0-dev
image: oci.stackable.tech/sdp/hadoop:3.4.0-stackable0.0.0-dev
# yamllint disable rule:line-length
command: ["bash", "-c", "/stackable/hadoop/bin/hdfs dfs -mkdir -p /ny-taxi-data/raw \
&& cd /tmp \
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: create-druid-ingestion-job
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor"]
volumeMounts:
- name: ingestion-job-spec
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: create-nifi-ingestion-job
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/IngestEarthquakesToKafka.xml && python -u /tmp/script/script.py"]
volumeMounts:
- name: script
Expand Down
2 changes: 1 addition & 1 deletion demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: setup-superset
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/superset-assets.zip && python -u /tmp/script/script.py"]
volumeMounts:
- name: script
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: create-druid-ingestion-job
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/stations-ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor && curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/measurements-ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor && curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/measurements-compaction-job-spec.json https://druid-coordinator:8281/druid/coordinator/v1/config/compaction"]
volumeMounts:
- name: ingestion-job-spec
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: create-nifi-ingestion-job
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/IngestWaterLevelsToKafka.xml && python -u /tmp/script/script.py"]
volumeMounts:
- name: script
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: setup-superset
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/superset-assets.zip && python -u /tmp/script/script.py"]
volumeMounts:
- name: script
Expand Down
2 changes: 1 addition & 1 deletion demos/signal-processing/Dockerfile-nifi
Original file line number Diff line number Diff line change
@@ -1,3 +1,3 @@
FROM docker.stackable.tech/stackable/nifi:1.27.0-stackable24.7.0
FROM oci.stackable.tech/sdp/nifi:1.27.0-stackable24.7.0

RUN curl --fail -o /stackable/nifi/postgresql-42.6.0.jar "https://repo.stackable.tech/repository/misc/postgresql-timescaledb/postgresql-42.6.0.jar"
4 changes: 2 additions & 2 deletions demos/signal-processing/create-nifi-ingestion-job.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -9,13 +9,13 @@ spec:
serviceAccountName: demo-serviceaccount
initContainers:
- name: wait-for-timescale-job
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0
image: oci.stackable.tech/sdp/tools:1.0.0-stackable24.7.0
command: ["bash", "-c", "echo 'Waiting for timescaleDB tables to be ready'
&& kubectl wait --for=condition=complete job/create-timescale-tables-job"
]
containers:
- name: create-nifi-ingestion-job
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "export PGPASSWORD=$(cat /timescale-admin-credentials/password) && \
curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/DownloadAndWriteToDB.xml && \
sed -i \"s/PLACEHOLDERPGPASSWORD/$PGPASSWORD/g\" DownloadAndWriteToDB.xml && \
Expand Down
2 changes: 1 addition & 1 deletion demos/signal-processing/create-timescale-tables.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ spec:
serviceAccountName: demo-serviceaccount
initContainers:
- name: wait-for-timescale
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0
image: oci.stackable.tech/sdp/tools:1.0.0-stackable24.7.0
command: ["bash", "-c", "echo 'Waiting for timescaleDB to be ready'
&& kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/name=postgresql-timescaledb"
]
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,11 +8,11 @@ spec:
spec:
initContainers:
- name: wait-for-testdata
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "echo 'Waiting for job load-ny-taxi-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-ny-taxi-data"]
containers:
- name: create-spark-anomaly-detection-job
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ad-job.yaml"]
volumeMounts:
- name: manifest
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: setup-superset
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"]
volumeMounts:
- name: script
Expand Down
2 changes: 1 addition & 1 deletion demos/trino-taxi-data/create-table-in-trino.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: create-ny-taxi-data-table-in-trino
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "python -u /tmp/script/script.py"]
volumeMounts:
- name: script
Expand Down
2 changes: 1 addition & 1 deletion demos/trino-taxi-data/setup-superset.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: setup-superset
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"]
volumeMounts:
- name: script
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -160,7 +160,7 @@ Libraries can be added to a custom *product* image launched by the notebook. Sup
spark = (SparkSession
.builder
.master(f'k8s://https://{os.environ["KUBERNETES_SERVICE_HOST"]}:{os.environ["KUBERNETES_SERVICE_PORT"]}')
.config("spark.kubernetes.container.image", "docker.stackable.tech/demos/spark-k8s-with-scikit-learn:3.5.0-stackable24.3.0")
.config("spark.kubernetes.container.image", "oci.stackable.tech/stackable/spark-k8s-with-scikit-learn:3.5.0-stackable24.3.0")
.config("spark.driver.port", "2222")
.config("spark.driver.blockManager.port", "7777")
.config("spark.driver.host", "driver-service.default.svc.cluster.local")
Expand All @@ -183,16 +183,16 @@ It requires a specific Spark image:
[source,python]
----
.config("spark.kubernetes.container.image",
"docker.stackable.tech/demos/spark-k8s-with-scikit-learn:3.5.0-stackable24.3.0"),
"oci.stackable.tech/stackable/spark-k8s-with-scikit-learn:3.5.0-stackable24.3.0"),
...
----

This is created by taking a Spark image, in this case `docker.stackable.tech/stackable/spark-k8s:3.5.0-stackable24.3.0`, installing specific python libraries into it
This is created by taking a Spark image, in this case `oci.stackable.tech/sdp/spark-k8s:3.5.0-stackable24.3.0`, installing specific python libraries into it
, and re-tagging the image:

[source,console]
----
FROM docker.stackable.tech/stackable/spark-k8s:3.5.0-stackable24.3.0
FROM oci.stackable.tech/sdp/spark-k8s:3.5.0-stackable24.3.0

COPY demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/requirements.txt .

Expand Down
2 changes: 1 addition & 1 deletion stacks/_templates/jupyterhub.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,7 @@ options:
HADOOP_CONF_DIR: "/home/jovyan/hdfs"
initContainers:
- name: download-notebook
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0
image: oci.stackable.tech/sdp/tools:1.0.0-stackable24.7.0
command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb -o /notebook/notebook.ipynb']
volumeMounts:
- mountPath: /notebook
Expand Down
2 changes: 1 addition & 1 deletion stacks/_templates/keycloak.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ spec:
- name: tls
mountPath: /tls/
- name: create-auth-class
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command: ["/bin/bash", "-c"]
args:
- |
Expand Down
8 changes: 4 additions & 4 deletions stacks/end-to-end-security/krb5.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@ spec:
spec:
initContainers:
- name: init
image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.7.0
image: oci.stackable.tech/sdp/krb5:1.21.1-stackable24.7.0
args:
- sh
- -euo
Expand All @@ -35,7 +35,7 @@ spec:
name: data
containers:
- name: kdc
image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.7.0
image: oci.stackable.tech/sdp/krb5:1.21.1-stackable24.7.0
args:
- krb5kdc
- -n
Expand All @@ -48,7 +48,7 @@ spec:
- mountPath: /var/kerberos/krb5kdc
name: data
- name: kadmind
image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.7.0
image: oci.stackable.tech/sdp/krb5:1.21.1-stackable24.7.0
args:
- kadmind
- -nofork
Expand All @@ -61,7 +61,7 @@ spec:
- mountPath: /var/kerberos/krb5kdc
name: data
- name: client
image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.7.0
image: oci.stackable.tech/sdp/krb5:1.21.1-stackable24.7.0
tty: true
stdin: true
env:
Expand Down
2 changes: 1 addition & 1 deletion stacks/end-to-end-security/superset.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ spec:
initContainers:
# The postgres image does not contain curl or wget...
- name: download-dump
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command:
- bash
- -c
Expand Down
2 changes: 1 addition & 1 deletion stacks/jupyterhub-pyspark-hdfs/notebook.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@
" SparkSession\n",
" .builder\n",
" .master(f'k8s://https://{os.environ[\"KUBERNETES_SERVICE_HOST\"]}:{os.environ[\"KUBERNETES_SERVICE_PORT\"]}')\n",
" .config(\"spark.kubernetes.container.image\", \"docker.stackable.tech/demos/spark-k8s-with-scikit-learn:3.5.0-stackable24.3.0\")\n",
" .config(\"spark.kubernetes.container.image\", \"oci.stackable.tech/stackable/spark-k8s-with-scikit-learn:3.5.0-stackable24.3.0\")\n",
" .config(\"spark.driver.port\", \"2222\")\n",
" .config(\"spark.driver.blockManager.port\", \"7777\")\n",
" .config(\"spark.driver.host\", f\"driver-service.{NAMESPACE}.svc.cluster.local\")\n",
Expand Down
2 changes: 1 addition & 1 deletion stacks/keycloak-opa-poc/keycloak.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@ spec:
spec:
containers:
- name: propagate-keycloak-address
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
command:
- bash
- -x
Expand Down
2 changes: 1 addition & 1 deletion stacks/keycloak-opa-poc/setup-keycloak.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@ spec:
spec:
containers:
- name: setup-keycloak
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
env:
- name: KEYCLOAK_ADMIN_PASSWORD
valueFrom:
Expand Down
2 changes: 1 addition & 1 deletion stacks/logging/setup-opensearch-dashboards.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ spec:
spec:
containers:
- name: setup-opensearch-dashboards
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0
image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable24.7.0
env:
- name: OPEN_SEARCH_ADMIN_PASSWORD
valueFrom:
Expand Down
Loading