Skip to content

Commit 23931df

Browse files
pull main
2 parents b0ca44f + 6445d5e commit 23931df

File tree

7 files changed

+100
-46
lines changed

7 files changed

+100
-46
lines changed

.github/assets-config.yml

Lines changed: 0 additions & 4 deletions
This file was deleted.

.github/auto-labeler-config.yaml

Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,23 @@
1+
---
2+
conventional-commits:
3+
- type: 'fix'
4+
nouns: ['FIX', 'Fix', 'fix', 'FIXED', 'Fixed', 'fixed']
5+
labels: ['bug']
6+
- type: 'feature'
7+
nouns: ['FEATURE', 'Feature', 'feature', 'FEAT', 'Feat', 'feat']
8+
labels: ['enhancement']
9+
- type: 'breaking_change'
10+
nouns: ['BREAKING CHANGE', 'BREAKING', 'MAJOR']
11+
labels: ['breaking-change']
12+
- type: 'refactor'
13+
nouns: ['refactor', 'Refactor']
14+
labels: ['refactor']
15+
- type: 'documentation'
16+
nouns: ['docs', 'doc', 'document', 'documentation']
17+
labels: ['documentation']
18+
- type: 'build'
19+
nouns: ['build', 'rebuild']
20+
labels: ['build']
21+
- type: 'config'
22+
nouns: ['config', 'conf', 'configuration', 'configure']
23+
labels: ['config']

.github/docker-build-config.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
{
1414
"name": "qubership-zookeeper-transfer",
1515
"file": "docker-transfer/Dockerfile",
16-
"context": ""
16+
"context": "."
1717
},
1818
{
1919
"name": "qubership-zookeeper-backup-daemon",

.github/workflows/helm-charts-release.yaml

Lines changed: 15 additions & 35 deletions
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,6 @@ jobs:
2929
check-tag: true
3030
env:
3131
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
32-
3332
load-docker-build-components:
3433
runs-on: ubuntu-latest
3534
outputs:
@@ -99,28 +98,18 @@ jobs:
9998
fetch-depth: 0
10099
- name: "Update versions in values"
101100
id: update-versions
102-
uses: netcracker/qubership-workflow-hub/actions/helm-charts-release@main
101+
uses: netcracker/qubership-workflow-hub/actions/charts-values-update-action@main
103102
with:
104103
release-version: ${{ inputs.release }}
105-
config-file: .github/helm-charts-release-config.yaml
104+
config-file: .github/charts-values-update-config.yaml
106105
env:
107106
${{ insert }}: ${{ vars }}
108107
- name: "Debug"
109108
run: |
110109
echo "Images versions: ${{ steps.update-versions.outputs.images-versions }}"
111-
tag:
112-
needs: [chart-release-prepare]
113-
runs-on: ubuntu-latest
114-
steps:
115-
- name: Create release tag
116-
uses: netcracker/qubership-workflow-hub/actions/tag-action@main
117-
with:
118-
tag-name: "${{ inputs.release }}"
119-
ref: "release-${{ inputs.release }}"
120-
create-tag: true
121-
check-tag: false
110+
122111
docker-build:
123-
needs: [tag, chart-release-prepare, load-docker-build-components]
112+
needs: [chart-release-prepare, load-docker-build-components]
124113
runs-on: ubuntu-latest
125114
strategy:
126115
fail-fast: true
@@ -135,18 +124,17 @@ jobs:
135124
- name: Docker build
136125
uses: netcracker/qubership-workflow-hub/actions/docker-action@main
137126
with:
138-
ref: ${{ inputs.release }}
127+
ref: release-${{ inputs.release }}
139128
download-artifact: false
140129
dry-run: false
141130
component: ${{ toJson(matrix.component) }}
142131
platforms: ${{ needs.load-docker-build-components.outputs.platforms }}
143132
tags: "${{ env.IMAGE_VERSION }},latest"
144133
env:
145134
GITHUB_TOKEN: ${{ github.token }}
146-
147135
charts-release:
148-
needs: [tag, docker-build]
149-
continue-on-error: true
136+
needs: [docker-build]
137+
continue-on-error: false
150138
runs-on: ubuntu-latest
151139
steps:
152140
- name: Checkout code
@@ -161,10 +149,13 @@ jobs:
161149
git config user.email "$GITHUB_ACTOR@users.noreply.github.com"
162150
163151
- name: Run chart-releaser
152+
#continue-on-error: true
164153
uses: netcracker/chart-releaser-action@main
165154
with:
166155
charts_dir: operator/charts/helm
167156
release_name_template: "{{ .Version }}"
157+
skip_existing: true
158+
skip_upload: true
168159
env:
169160
CR_TOKEN: "${{ secrets.GITHUB_TOKEN }}"
170161

@@ -178,21 +169,10 @@ jobs:
178169
version: ${{ inputs.release }}
179170
env:
180171
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
181-
upload-assets:
182-
needs: [charts-release]
183-
runs-on: ubuntu-latest
184-
steps:
185-
- name: Checkout code
186-
uses: actions/checkout@v4
172+
- name: Upload Assets
173+
uses: netcracker/qubership-workflow-hub/actions/assets-action@main
187174
with:
188-
ref: ${{ inputs.release }}
189-
190-
- name: Archive and Upload Assets
191-
uses: netcracker/qubership-workflow-hub/actions/archive-and-upload-assets@main
192-
with:
193-
config-path: './.github/assets-config.yml'
194-
dist-path: './dist'
195-
upload: true
196-
ref: ${{ inputs.release }}
175+
tag: ${{ inputs.release }}
176+
item-path: .cr-release-packages/*.tgz
197177
env:
198-
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
178+
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}

docs/public/installation.md

Lines changed: 59 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -894,13 +894,68 @@ availability zones.
894894
You can manage pods' distribution using `affinity` rules to prevent Kubernetes from running ZooKeeper pods on nodes of
895895
the same availability zone.
896896

897+
#### Affinity cross Kubernetes Nodes
898+
899+
Default affinity cross kubernetes Nodes (1 pod per 1 node).
900+
901+
<details>
902+
<summary>Click to expand YAML</summary>
903+
904+
```yaml
905+
zooKeeper:
906+
affinity: {
907+
"podAntiAffinity": {
908+
"requiredDuringSchedulingIgnoredDuringExecution": [
909+
{
910+
"labelSelector": {
911+
"matchExpressions": [
912+
{
913+
"key": "component",
914+
"operator": "In",
915+
"values": [
916+
"zookeeper"
917+
]
918+
}
919+
]
920+
},
921+
"topologyKey": "kubernetes.io/hostname"
922+
}
923+
]
924+
},
925+
"nodeAffinity": {
926+
"requiredDuringSchedulingIgnoredDuringExecution": {
927+
"nodeSelectorTerms": [
928+
{
929+
"matchExpressions": [
930+
{
931+
"key": "role",
932+
"operator": "In",
933+
"values": [
934+
"compute"
935+
]
936+
}
937+
]
938+
}
939+
]
940+
}
941+
}
942+
}
943+
```
944+
945+
</details>
946+
947+
Where:
948+
949+
* `kubernetes.io/hostname` is the name of the label that defines the node name. This is the default name for Kubernetes.
950+
* `role` and `compute` are the sample name and value of label that defines the region to run ZooKeeper pods.
951+
897952
**Note**: This section describes deployment only for `storage class` Persistent Volumes (PV) type because with
898953
Predefined PV, the ZooKeeper pods are started on the nodes that are specified explicitly with Persistent Volumes. In
899954
that way, it is necessary to take care of creating PVs on nodes belonging to different AZ in advance.
900955

901956
#### Replicas Fewer Than Availability Zones
902957

903-
For cases when the number of ZooKeeper pods (value of the `zookeeper.replicas` parameter) is equal or less than the number of
958+
For cases when the number of ZooKeeper pods (value of the `zooKeeper.replicas` parameter) is equal or less than the number of
904959
availability zones, you need to restrict the start of pods to one pod per availability zone.
905960
You can also specify additional node affinity rule to start pods on allowed Kubernetes nodes.
906961

@@ -910,7 +965,7 @@ For this, you can use the following affinity rules:
910965
<summary>Click to expand YAML</summary>
911966

912967
```yaml
913-
zookeeper:
968+
zooKeeper:
914969
affinity: {
915970
"podAntiAffinity": {
916971
"requiredDuringSchedulingIgnoredDuringExecution": [
@@ -960,7 +1015,7 @@ Where:
9601015

9611016
#### Replicas More Than Availability Zones
9621017

963-
For cases when the number of ZooKeeper pods (value of the `zookeeper.replicas` parameter) is greater than the number of availability
1018+
For cases when the number of ZooKeeper pods (value of the `zooKeeper.replicas` parameter) is greater than the number of availability
9641019
zones, you need to restrict the start of pods to one pod per node and specify the preferred rule to start on different
9651020
availability zones.
9661021
You can also specify an additional node affinity rule to start the pods on allowed Kubernetes nodes.
@@ -971,7 +1026,7 @@ For this, you can use the following affinity rules:
9711026
<summary>Click to expand YAML</summary>
9721027

9731028
```yaml
974-
zookeeper:
1029+
zooKeeper:
9751030
affinity: {
9761031
"podAntiAffinity": {
9771032
"requiredDuringSchedulingIgnoredDuringExecution": [

monitoring/docker/Dockerfile

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
FROM telegraf:1.35.1-alpine
1+
FROM telegraf:1.35.2-alpine
22

33
ENV ZOOKEEPER_MONITORING_HOME=/opt/zookeeper-monitoring
44

@@ -16,7 +16,7 @@ RUN rm /usr/lib/python3.12/EXTERNALLY-MANAGED
1616

1717
RUN python3 -m ensurepip \
1818
&& rm -r /usr/lib/python*/ensurepip \
19-
&& pip3 install --upgrade pip setuptools==70.0.0 \
19+
&& pip3 install --upgrade pip setuptools==80.9.0 \
2020
&& pip3 install -r ${ZOOKEEPER_MONITORING_HOME}/requirements.txt \
2121
&& chmod -R 777 ${ZOOKEEPER_MONITORING_HOME}/exec-scripts \
2222
&& chmod +x /docker-entrypoint.sh \

0 commit comments

Comments
 (0)