Skip to content

Commit ceb2125

Browse files
committed
Merge remote-tracking branch 'origin/main' into feat/opensearch
2 parents 61d50cf + 7d630e5 commit ceb2125

File tree

62 files changed

+588
-322
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

62 files changed

+588
-322
lines changed

.github/workflows/build_hello-world.yaml

Lines changed: 0 additions & 31 deletions
This file was deleted.

.scripts/update_readme_badges.sh

Lines changed: 9 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -54,12 +54,15 @@ for BUILD_WORKFLOW_FILE in .github/workflows/build_*.yaml; do
5454
echo >> "$BADGES_TMP"
5555
fi
5656
done
57-
# This needs to add the remaning empty columns of the last row in the table
58-
# This is a hack to fix the status quo and make markdownlint happy.
59-
for _ in $(seq 0 $((COLS - 1))); do
60-
echo -n "| " >> "$BADGES_TMP"
61-
done
62-
echo "|" >> "$BADGES_TMP"
57+
58+
# Add remaining empty columns to complete the last row if needed
59+
# "if needed" is the first if here: It'll only run when we're NOT on the last column (0 indexed)
60+
if [ ${CURRENT_COLUMN} -ne $((COLS - 1)) ]; then
61+
for _ in $(seq $((CURRENT_COLUMN + 1)) $((COLS - 1))); do
62+
echo -n "| " >> "$BADGES_TMP"
63+
done
64+
echo "|" >> "$BADGES_TMP"
65+
fi
6366
echo -n "<!-- end:badges -->" >> "$BADGES_TMP"
6467

6568
# Print the image and link shortcuts. Eg:

CHANGELOG.md

Lines changed: 8 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -59,6 +59,7 @@ All notable changes to this project will be documented in this file.
5959
- zookeeper: bump jetty version for CVE-2024-13009 in 3.9.3 ([#1179])
6060
- zookeeper: bump netty version for CVE-2025-24970 in 3.9.3 ([#1180])
6161
- hadoop: backport HADOOP-19352, HADOOP-19335, HADOOP-19465, HADOOP-19456 and HADOOP-19225 to fix vulnerabilities in Hadoop `3.4.1` ([#1184])
62+
- hadoop: Backport HADOOP-18583 to make OpenSSL 3.x work with the native hadoop libraries ([#1209]).
6263
- opensearch: Add Opensearch as new product with version `3.1.0` ([#1215]).
6364

6465
### Changed
@@ -93,7 +94,10 @@ All notable changes to this project will be documented in this file.
9394
- opa: Enable custom versions ([#1170]).
9495
- use custom product versions for Hadoop, HBase, Phoenix, hbase-operator-tools, Druid, Hive and Spark ([#1173]).
9596
- hbase: Bump dependencies to the latest patch level for HBase `2.6.1` and `2.6.2` ([#1185]).
96-
- Changed default user & group IDs from 1000/1000 to 782252253/574654813 ([#1164])
97+
- hadoop: Separate Dockerfiles for Hadoop build and HDFS image ([#1186]).
98+
- ubi-rust-builder: Bump Rust toolchain to 1.87.0, cargo-auditable to 0.7.0 and protoc to 31.1 ([#1197]).
99+
- stackable-base, stackable-devel, ubi-rust-builder: Update `ubi-minimal` base image ([#1197]).
100+
- testing-tools: Update `python` 3.12-slim-bullseye base image ([#1197]).
97101

98102
### Fixed
99103

@@ -203,7 +207,6 @@ All notable changes to this project will be documented in this file.
203207
[#1156]: https://github.com/stackabletech/docker-images/pull/1156
204208
[#1159]: https://github.com/stackabletech/docker-images/pull/1159
205209
[#1163]: https://github.com/stackabletech/docker-images/pull/1163
206-
[#1164]: https://github.com/stackabletech/docker-images/pull/1164
207210
[#1165]: https://github.com/stackabletech/docker-images/pull/1165
208211
[#1168]: https://github.com/stackabletech/docker-images/pull/1168
209212
[#1169]: https://github.com/stackabletech/docker-images/pull/1169
@@ -216,8 +219,11 @@ All notable changes to this project will be documented in this file.
216219
[#1180]: https://github.com/stackabletech/docker-images/pull/1180
217220
[#1184]: https://github.com/stackabletech/docker-images/pull/1184
218221
[#1185]: https://github.com/stackabletech/docker-images/pull/1185
222+
[#1186]: https://github.com/stackabletech/docker-images/pull/1186
219223
[#1188]: https://github.com/stackabletech/docker-images/pull/1188
220224
[#1189]: https://github.com/stackabletech/docker-images/pull/1189
225+
[#1197]: https://github.com/stackabletech/docker-images/pull/1197
226+
[#1209]: https://github.com/stackabletech/docker-images/pull/1209
221227
[#1215]: https://github.com/stackabletech/docker-images/pull/1215
222228

223229
## [25.3.0] - 2025-03-21

README.md

Lines changed: 5 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -6,12 +6,11 @@ This repository contains Dockerfiles and scripts to build base images for use wi
66
| | | | |
77
| -: | -: | -: | -: |
88
| [![Build Airflow]][build_airflow.yaml] | [![Build Druid]][build_druid.yaml] | [![Build Hadoop]][build_hadoop.yaml] | [![Build HBase]][build_hbase.yaml] |
9-
| [![Build Hello-World]][build_hello-world.yaml] | [![Build Hive]][build_hive.yaml] | [![Build Java Base]][build_java-base.yaml] | [![Build Java Development]][build_java-devel.yaml] |
10-
| [![Build Kafka Testing Tools]][build_kafka-testing-tools.yaml] | [![Build Kafka]][build_kafka.yaml] | [![Build Krb5]][build_krb5.yaml] | [![Build NiFi]][build_nifi.yaml] |
11-
| [![Build Omid]][build_omid.yaml] | [![Build OPA]][build_opa.yaml] | [![Build Spark Connect Client]][build_spark-connect-client.yaml] | [![Build Spark K8s]][build_spark-k8s.yaml] |
12-
| [![Build Stackable Base]][build_stackable-base.yaml] | [![Build Superset]][build_superset.yaml] | [![Build Testing Tools]][build_testing-tools.yaml] | [![Build Tools]][build_tools.yaml] |
13-
| [![Build Trino CLI]][build_trino-cli.yaml] | [![Build Trino]][build_trino.yaml] | [![Build Vector]][build_vector.yaml] | [![Build ZooKeeper]][build_zookeeper.yaml] |
14-
| | | | |
9+
| [![Build Hive]][build_hive.yaml] | [![Build Java Base]][build_java-base.yaml] | [![Build Java Development]][build_java-devel.yaml] | [![Build Kafka Testing Tools]][build_kafka-testing-tools.yaml] |
10+
| [![Build Kafka]][build_kafka.yaml] | [![Build Krb5]][build_krb5.yaml] | [![Build NiFi]][build_nifi.yaml] | [![Build Omid]][build_omid.yaml] |
11+
| [![Build OPA]][build_opa.yaml] | [![Build Spark Connect Client]][build_spark-connect-client.yaml] | [![Build Spark K8s]][build_spark-k8s.yaml] | [![Build Stackable Base]][build_stackable-base.yaml] |
12+
| [![Build Superset]][build_superset.yaml] | [![Build Testing Tools]][build_testing-tools.yaml] | [![Build Tools]][build_tools.yaml] | [![Build Trino CLI]][build_trino-cli.yaml] |
13+
| [![Build Trino]][build_trino.yaml] | [![Build Vector]][build_vector.yaml] | [![Build ZooKeeper]][build_zookeeper.yaml] | |
1514
<!-- end:badges -->
1615

1716
## Prerequisites
@@ -222,8 +221,6 @@ ENTRYPOINT ["/stackable-zookeeper-operator"]
222221
[build_hadoop.yaml]: https://github.com/stackabletech/docker-images/actions/workflows/build_hadoop.yaml
223222
[Build HBase]: https://github.com/stackabletech/docker-images/actions/workflows/build_hbase.yaml/badge.svg
224223
[build_hbase.yaml]: https://github.com/stackabletech/docker-images/actions/workflows/build_hbase.yaml
225-
[Build Hello-World]: https://github.com/stackabletech/docker-images/actions/workflows/build_hello-world.yaml/badge.svg
226-
[build_hello-world.yaml]: https://github.com/stackabletech/docker-images/actions/workflows/build_hello-world.yaml
227224
[Build Hive]: https://github.com/stackabletech/docker-images/actions/workflows/build_hive.yaml/badge.svg
228225
[build_hive.yaml]: https://github.com/stackabletech/docker-images/actions/workflows/build_hive.yaml
229226
[Build Java Base]: https://github.com/stackabletech/docker-images/actions/workflows/build_java-base.yaml/badge.svg

conf.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -13,12 +13,12 @@
1313
airflow = importlib.import_module("airflow.versions")
1414
druid = importlib.import_module("druid.versions")
1515
hadoop = importlib.import_module("hadoop.versions")
16+
hadoop_jars = importlib.import_module("hadoop.hadoop.versions")
1617
hbase = importlib.import_module("hbase.versions")
1718
hbase_jars = importlib.import_module("hbase.hbase.versions")
1819
hbase_phoenix = importlib.import_module("hbase.phoenix.versions")
1920
hbase_opa_authorizer = importlib.import_module("hbase.hbase-opa-authorizer.versions")
2021
hbase_operator_tools = importlib.import_module("hbase.hbase-operator-tools.versions")
21-
hello_world = importlib.import_module("hello-world.versions")
2222
hive = importlib.import_module("hive.versions")
2323
java_base = importlib.import_module("java-base.versions")
2424
java_devel = importlib.import_module("java-devel.versions")
@@ -54,12 +54,12 @@
5454
{"name": "airflow", "versions": airflow.versions},
5555
{"name": "druid", "versions": druid.versions},
5656
{"name": "hadoop", "versions": hadoop.versions},
57+
{"name": "hadoop/hadoop", "versions": hadoop_jars.versions},
5758
{"name": "hbase", "versions": hbase.versions},
5859
{"name": "hbase/hbase", "versions": hbase_jars.versions},
5960
{"name": "hbase/phoenix", "versions": hbase_phoenix.versions},
6061
{"name": "hbase/hbase-opa-authorizer", "versions": hbase_opa_authorizer.versions},
6162
{"name": "hbase/hbase-operator-tools", "versions": hbase_operator_tools.versions},
62-
{"name": "hello-world", "versions": hello_world.versions},
6363
{"name": "hive", "versions": hive.versions},
6464
{"name": "java-base", "versions": java_base.versions},
6565
{"name": "java-devel", "versions": java_devel.versions},
@@ -121,7 +121,7 @@
121121

122122
args = {
123123
"STACKABLE_USER_NAME": "stackable",
124-
"STACKABLE_USER_UID": "782252253", # This is a random high id to not conflict with any existing user
125-
"STACKABLE_USER_GID": "574654813", # This is a random high id to not conflict with any existing group
124+
"STACKABLE_USER_UID": "1000",
125+
"STACKABLE_USER_GID": "1000",
126126
"DELETE_CACHES": "true",
127127
}

druid/Dockerfile

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
# syntax=docker/dockerfile:1.16.0@sha256:e2dd261f92e4b763d789984f6eab84be66ab4f5f08052316d8eb8f173593acf7
22
# check=error=true
33

4-
FROM stackable/image/hadoop AS hadoop-builder
4+
FROM stackable/image/hadoop/hadoop AS hadoop-builder
55

66
FROM stackable/image/java-devel AS druid-builder
77

@@ -12,7 +12,9 @@ ARG STAX2_API
1212
ARG WOODSTOX_CORE
1313
ARG AUTHORIZER
1414
ARG STACKABLE_USER_UID
15-
ARG HADOOP
15+
ARG HADOOP_HADOOP
16+
# Reassign the arg to `HADOOP_VERSION` for better readability.
17+
ENV HADOOP_VERSION=${HADOOP_HADOOP}
1618

1719
# Setting this to anything other than "true" will keep the cache folders around (e.g. for Maven, NPM etc.)
1820
# This can be used to speed up builds when disk space is of no concern.
@@ -75,7 +77,7 @@ mvn \
7577
--no-transfer-progress \
7678
clean install \
7779
-Pdist,stackable-bundle-contrib-exts \
78-
-Dhadoop.compile.version=${HADOOP}-stackable${RELEASE} \
80+
-Dhadoop.compile.version=${HADOOP_VERSION}-stackable${RELEASE} \
7981
-DskipTests `# Skip test execution` \
8082
-Dcheckstyle.skip `# Skip checkstyle checks. We dont care if the code is properly formatted, it just wastes time` \
8183
-Dmaven.javadoc.skip=true `# Dont generate javadoc` \

druid/versions.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -4,23 +4,23 @@
44
# https://druid.apache.org/docs/30.0.1/operations/java/
55
"java-base": "17",
66
"java-devel": "17",
7-
"hadoop": "3.3.6",
7+
"hadoop/hadoop": "3.3.6",
88
"authorizer": "0.7.0",
99
},
1010
{
1111
"product": "31.0.1",
1212
# https://druid.apache.org/docs/31.0.1/operations/java/
1313
"java-base": "17",
1414
"java-devel": "17",
15-
"hadoop": "3.3.6",
15+
"hadoop/hadoop": "3.3.6",
1616
"authorizer": "0.7.0",
1717
},
1818
{
1919
"product": "33.0.0",
2020
# https://druid.apache.org/docs/33.0.0/operations/java/
2121
"java-base": "17",
2222
"java-devel": "17",
23-
"hadoop": "3.3.6",
23+
"hadoop/hadoop": "3.3.6",
2424
"authorizer": "0.7.0",
2525
},
2626
]

0 commit comments

Comments
 (0)