diff --git a/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml b/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml index efa281b2..f61a9aad 100644 --- a/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml +++ b/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: start-pyspark-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/tools:1.0.0-stackable0.0.0-dev # N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly # restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps # below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that diff --git a/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml b/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml index f74bab50..d72fe62c 100644 --- a/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml +++ b/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: start-date-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/tools:1.0.0-stackable0.0.0-dev # N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly # restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps # below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml index 6bcc924d..a33c2973 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml @@ -9,11 +9,11 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-kafka - image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/tools:1.0.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/instance=kafka -l app.kubernetes.io/name=kafka"] containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/LakehouseKafkaIngest.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml index db28afc8..6ef43a43 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml @@ -12,11 +12,11 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-kafka - image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/tools:1.0.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/name=kafka -l app.kubernetes.io/instance=kafka"] containers: - name: create-spark-ingestion-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/tools:1.0.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ingestion-job.yaml"] volumeMounts: - name: manifest diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml index 6c0317bb..5a0f5ae9 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml @@ -9,11 +9,11 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-testdata - image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/tools:1.0.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Waiting for job load-test-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-test-data"] containers: - name: create-tables-in-trino - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml b/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml index 1f70f900..3275c927 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/end-to-end-security/create-spark-report.yaml b/demos/end-to-end-security/create-spark-report.yaml index 2e1bff2d..63829826 100644 --- a/demos/end-to-end-security/create-spark-report.yaml +++ b/demos/end-to-end-security/create-spark-report.yaml @@ -12,7 +12,7 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-trino-tables - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: - bash - -euo @@ -23,7 +23,7 @@ spec: kubectl wait --timeout=30m --for=condition=complete job/create-tables-in-trino containers: - name: create-spark-report - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: - bash - -euo diff --git a/demos/end-to-end-security/create-trino-tables.yaml b/demos/end-to-end-security/create-trino-tables.yaml index 16b0e6f1..8778acb5 100644 --- a/demos/end-to-end-security/create-trino-tables.yaml +++ b/demos/end-to-end-security/create-trino-tables.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-tables-in-trino - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml b/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml index d250e0db..78e03d84 100644 --- a/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml +++ b/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml @@ -9,7 +9,7 @@ spec: spec: containers: - name: create-hfile-and-import-to-hbase - image: docker.stackable.tech/stackable/hbase:2.4.18-stackable0.0.0-dev + image: oci.stackable.tech/sdp/hbase:2.4.18-stackable0.0.0-dev env: - name: HADOOP_USER_NAME value: stackable diff --git a/demos/hbase-hdfs-load-cycling-data/distcp-cycling-data.yaml b/demos/hbase-hdfs-load-cycling-data/distcp-cycling-data.yaml index ad564e6f..eed72029 100644 --- a/demos/hbase-hdfs-load-cycling-data/distcp-cycling-data.yaml +++ b/demos/hbase-hdfs-load-cycling-data/distcp-cycling-data.yaml @@ -11,7 +11,7 @@ spec: # We use 24.3.0 here which contains the distcp MapReduce components # This is not included in the 24.7 and 24.11 images and will fail. # See: https://github.com/stackabletech/docker-images/issues/793 - image: docker.stackable.tech/stackable/hadoop:3.3.6-stackable24.3.0 + image: oci.stackable.tech/sdp/hadoop:3.3.6-stackable24.3.0 env: - name: HADOOP_USER_NAME value: stackable diff --git a/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/Dockerfile b/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/Dockerfile index 232b7c8b..64a7821e 100644 --- a/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/Dockerfile +++ b/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/Dockerfile @@ -1,4 +1,4 @@ -FROM docker.stackable.tech/stackable/spark-k8s:3.5.0-stackable24.3.0 +FROM oci.stackable.tech/sdp/spark-k8s:3.5.0-stackable24.3.0 COPY demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/requirements.txt . diff --git a/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml b/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml index d02c5088..ecfaa50b 100644 --- a/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml +++ b/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: load-ny-taxi-data - image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/hadoop:3.4.0-stackable0.0.0-dev # yamllint disable rule:line-length command: ["bash", "-c", "/stackable/hadoop/bin/hdfs dfs -mkdir -p /ny-taxi-data/raw \ && cd /tmp \ diff --git a/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml b/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml index 93bedf98..421ba348 100644 --- a/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-druid-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor"] volumeMounts: - name: ingestion-job-spec diff --git a/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml b/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml index 0d1aa1cc..19bb7675 100644 --- a/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/IngestEarthquakesToKafka.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml b/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml index 36bc2110..15ee14c9 100644 --- a/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml b/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml index 88b8fea5..01570f29 100644 --- a/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-druid-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/stations-ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor && curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/measurements-ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor && curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/measurements-compaction-job-spec.json https://druid-coordinator:8281/druid/coordinator/v1/config/compaction"] volumeMounts: - name: ingestion-job-spec diff --git a/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml b/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml index ff809981..7dcf70ed 100644 --- a/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/IngestWaterLevelsToKafka.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml b/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml index 66f3ef56..fad556f2 100644 --- a/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml +++ b/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/signal-processing/Dockerfile-nifi b/demos/signal-processing/Dockerfile-nifi index eb242ad3..b53549da 100644 --- a/demos/signal-processing/Dockerfile-nifi +++ b/demos/signal-processing/Dockerfile-nifi @@ -1,3 +1,3 @@ -FROM docker.stackable.tech/stackable/nifi:1.27.0-stackable0.0.0-dev +FROM oci.stackable.tech/sdp/nifi:1.27.0-stackable0.0.0-dev RUN curl --fail -o /stackable/nifi/postgresql-42.6.0.jar "https://repo.stackable.tech/repository/misc/postgresql-timescaledb/postgresql-42.6.0.jar" diff --git a/demos/signal-processing/create-nifi-ingestion-job.yaml b/demos/signal-processing/create-nifi-ingestion-job.yaml index f63712bc..4b44a343 100644 --- a/demos/signal-processing/create-nifi-ingestion-job.yaml +++ b/demos/signal-processing/create-nifi-ingestion-job.yaml @@ -9,13 +9,13 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-timescale-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/tools:1.0.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Waiting for timescaleDB tables to be ready' && kubectl wait --for=condition=complete job/create-timescale-tables-job" ] containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "export PGPASSWORD=$(cat /timescale-admin-credentials/password) && \ curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/DownloadAndWriteToDB.xml && \ sed -i \"s/PLACEHOLDERPGPASSWORD/$PGPASSWORD/g\" DownloadAndWriteToDB.xml && \ diff --git a/demos/signal-processing/create-timescale-tables.yaml b/demos/signal-processing/create-timescale-tables.yaml index 173639b0..853ea310 100644 --- a/demos/signal-processing/create-timescale-tables.yaml +++ b/demos/signal-processing/create-timescale-tables.yaml @@ -9,7 +9,7 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-timescale - image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/tools:1.0.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Waiting for timescaleDB to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/name=postgresql-timescaledb" ] diff --git a/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml b/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml index a4b46e46..03516a3a 100644 --- a/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml +++ b/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml @@ -8,11 +8,11 @@ spec: spec: initContainers: - name: wait-for-testdata - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Waiting for job load-ny-taxi-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-ny-taxi-data"] containers: - name: create-spark-anomaly-detection-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ad-job.yaml"] volumeMounts: - name: manifest diff --git a/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml b/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml index 8a1a8cbc..56f8d5a3 100644 --- a/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml +++ b/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/trino-taxi-data/create-table-in-trino.yaml b/demos/trino-taxi-data/create-table-in-trino.yaml index 44e2500e..de3fc97e 100644 --- a/demos/trino-taxi-data/create-table-in-trino.yaml +++ b/demos/trino-taxi-data/create-table-in-trino.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-ny-taxi-data-table-in-trino - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/trino-taxi-data/setup-superset.yaml b/demos/trino-taxi-data/setup-superset.yaml index 792b459b..cb613aec 100644 --- a/demos/trino-taxi-data/setup-superset.yaml +++ b/demos/trino-taxi-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/docs/modules/demos/pages/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data.adoc b/docs/modules/demos/pages/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data.adoc index 10f04d62..7bc3cb9d 100644 --- a/docs/modules/demos/pages/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data.adoc +++ b/docs/modules/demos/pages/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data.adoc @@ -159,7 +159,7 @@ Libraries can be added to a custom *product* image launched by the notebook. Sup spark = (SparkSession .builder .master(f'k8s://https://{os.environ["KUBERNETES_SERVICE_HOST"]}:{os.environ["KUBERNETES_SERVICE_PORT"]}') - .config("spark.kubernetes.container.image", "docker.stackable.tech/demos/spark-k8s-with-scikit-learn:3.5.0-stackable24.3.0") + .config("spark.kubernetes.container.image", "oci.stackable.tech/stackable/spark-k8s-with-scikit-learn:3.5.0-stackable24.3.0") .config("spark.driver.port", "2222") .config("spark.driver.blockManager.port", "7777") .config("spark.driver.host", "driver-service.default.svc.cluster.local") @@ -182,16 +182,16 @@ It requires a specific Spark image: [source,python] ---- .config("spark.kubernetes.container.image", - "docker.stackable.tech/demos/spark-k8s-with-scikit-learn:3.5.0-stackable24.3.0"), + "oci.stackable.tech/stackable/spark-k8s-with-scikit-learn:3.5.0-stackable24.3.0"), ... ---- -This is created by taking a Spark image, in this case `docker.stackable.tech/stackable/spark-k8s:3.5.0-stackable24.3.0`, installing specific python libraries into it +This is created by taking a Spark image, in this case `oci.stackable.tech/sdp/spark-k8s:3.5.0-stackable24.3.0`, installing specific python libraries into it , and re-tagging the image: [source,console] ---- -FROM docker.stackable.tech/stackable/spark-k8s:3.5.0-stackable24.3.0 +FROM oci.stackable.tech/sdp/spark-k8s:3.5.0-stackable24.3.0 COPY demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/requirements.txt . diff --git a/stacks/_templates/jupyterhub.yaml b/stacks/_templates/jupyterhub.yaml index 05a28654..a5d1c0d2 100644 --- a/stacks/_templates/jupyterhub.yaml +++ b/stacks/_templates/jupyterhub.yaml @@ -50,7 +50,7 @@ options: HADOOP_CONF_DIR: "/home/jovyan/hdfs" initContainers: - name: download-notebook - image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/tools:1.0.0-stackable0.0.0-dev command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb -o /notebook/notebook.ipynb'] volumeMounts: - mountPath: /notebook diff --git a/stacks/_templates/keycloak.yaml b/stacks/_templates/keycloak.yaml index 345cf15a..9ebaa800 100644 --- a/stacks/_templates/keycloak.yaml +++ b/stacks/_templates/keycloak.yaml @@ -48,7 +48,7 @@ spec: - name: tls mountPath: /tls/ - name: create-auth-class - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: ["/bin/bash", "-c"] args: - | diff --git a/stacks/end-to-end-security/krb5.yaml b/stacks/end-to-end-security/krb5.yaml index 3ce20bd4..e92aa69f 100644 --- a/stacks/end-to-end-security/krb5.yaml +++ b/stacks/end-to-end-security/krb5.yaml @@ -14,7 +14,7 @@ spec: spec: initContainers: - name: init - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable0.0.0-dev + image: oci.stackable.tech/sdp/krb5:1.21.1-stackable0.0.0-dev args: - sh - -euo @@ -35,7 +35,7 @@ spec: name: data containers: - name: kdc - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable0.0.0-dev + image: oci.stackable.tech/sdp/krb5:1.21.1-stackable0.0.0-dev args: - krb5kdc - -n @@ -48,7 +48,7 @@ spec: - mountPath: /var/kerberos/krb5kdc name: data - name: kadmind - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable0.0.0-dev + image: oci.stackable.tech/sdp/krb5:1.21.1-stackable0.0.0-dev args: - kadmind - -nofork @@ -61,7 +61,7 @@ spec: - mountPath: /var/kerberos/krb5kdc name: data - name: client - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable0.0.0-dev + image: oci.stackable.tech/sdp/krb5:1.21.1-stackable0.0.0-dev tty: true stdin: true env: diff --git a/stacks/end-to-end-security/superset.yaml b/stacks/end-to-end-security/superset.yaml index c606075c..5ea9f460 100644 --- a/stacks/end-to-end-security/superset.yaml +++ b/stacks/end-to-end-security/superset.yaml @@ -25,7 +25,7 @@ spec: initContainers: # The postgres image does not contain curl or wget... - name: download-dump - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: - bash - -c diff --git a/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb b/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb index 2e839540..3058c6a0 100644 --- a/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb +++ b/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb @@ -45,7 +45,7 @@ " SparkSession\n", " .builder\n", " .master(f'k8s://https://{os.environ[\"KUBERNETES_SERVICE_HOST\"]}:{os.environ[\"KUBERNETES_SERVICE_PORT\"]}')\n", - " .config(\"spark.kubernetes.container.image\", \"docker.stackable.tech/demos/spark-k8s-with-scikit-learn:3.5.0-stackable24.3.0\")\n", + " .config(\"spark.kubernetes.container.image\", \"oci.stackable.tech/stackable/spark-k8s-with-scikit-learn:3.5.0-stackable24.3.0\")\n", " .config(\"spark.driver.port\", \"2222\")\n", " .config(\"spark.driver.blockManager.port\", \"7777\")\n", " .config(\"spark.driver.host\", f\"driver-service.{NAMESPACE}.svc.cluster.local\")\n", diff --git a/stacks/keycloak-opa-poc/keycloak.yaml b/stacks/keycloak-opa-poc/keycloak.yaml index eaf5d60b..a5d64637 100644 --- a/stacks/keycloak-opa-poc/keycloak.yaml +++ b/stacks/keycloak-opa-poc/keycloak.yaml @@ -70,7 +70,7 @@ spec: spec: containers: - name: propagate-keycloak-address - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev command: - bash - -x diff --git a/stacks/keycloak-opa-poc/setup-keycloak.yaml b/stacks/keycloak-opa-poc/setup-keycloak.yaml index ceba42fe..fca8cdd5 100644 --- a/stacks/keycloak-opa-poc/setup-keycloak.yaml +++ b/stacks/keycloak-opa-poc/setup-keycloak.yaml @@ -29,7 +29,7 @@ spec: spec: containers: - name: setup-keycloak - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev env: - name: KEYCLOAK_ADMIN_PASSWORD valueFrom: diff --git a/stacks/logging/setup-opensearch-dashboards.yaml b/stacks/logging/setup-opensearch-dashboards.yaml index b7e00641..bf707449 100644 --- a/stacks/logging/setup-opensearch-dashboards.yaml +++ b/stacks/logging/setup-opensearch-dashboards.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-opensearch-dashboards - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/testing-tools:0.2.0-stackable0.0.0-dev env: - name: OPEN_SEARCH_ADMIN_PASSWORD valueFrom: diff --git a/stacks/signal-processing/jupyterhub.yaml b/stacks/signal-processing/jupyterhub.yaml index f9782df5..58240d45 100644 --- a/stacks/signal-processing/jupyterhub.yaml +++ b/stacks/signal-processing/jupyterhub.yaml @@ -32,7 +32,7 @@ options: image: # TODO (@NickLarsenNZ): Use a versioned image with stackable0.0.0-dev or stackableXX.X.X so that # the demo is reproducable for the release and it will be automatically replaced for the release branch. - name: docker.stackable.tech/demos/jupyter-pyspark-with-alibi-detect + name: oci.stackable.tech/stackable/jupyter-pyspark-with-alibi-detect tag: python-3.9 serviceAccountName: spark networkPolicy: @@ -41,7 +41,7 @@ options: stackable.tech/vendor: Stackable initContainers: - name: download-notebook - image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev + image: oci.stackable.tech/sdp/tools:1.0.0-stackable0.0.0-dev command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/tsdb.ipynb -o /notebook/tsdb.ipynb'] volumeMounts: - mountPath: /notebook diff --git a/stacks/signal-processing/nifi.yaml b/stacks/signal-processing/nifi.yaml index 48d7c39d..7f3d3964 100644 --- a/stacks/signal-processing/nifi.yaml +++ b/stacks/signal-processing/nifi.yaml @@ -8,7 +8,7 @@ spec: productVersion: 1.27.0 # TODO (@NickLarsenNZ): Use a versioned image with stackable0.0.0-dev or stackableXX.X.X so that # the demo is reproducable for the release and it will be automatically replaced for the release branch. - custom: docker.stackable.tech/demos/nifi:1.27.0-postgresql + custom: oci.stackable.tech/stackable/nifi:1.27.0-postgresql clusterConfig: listenerClass: external-unstable zookeeperConfigMapName: nifi-znode