Skip to content

Commit 1c83e06

Browse files
Merge pull request #286895 from vnikolin/vnikolin-hwv3
HWV Updates for Release 3.13
2 parents b12e3e3 + 55eed8c commit 1c83e06

File tree

3 files changed

+97
-74
lines changed

3 files changed

+97
-74
lines changed

articles/operator-nexus/howto-configure-cluster.md

Lines changed: 21 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -86,12 +86,12 @@ az networkcloud cluster create --name "$CLUSTER_NAME" --location "$LOCATION" \
8686
| COMPX_RACK_SKU | Rack SKU for CompX Rack; repeat for each rack in compute-rack-definitions |
8787
| COMPX_RACK_SN | Rack Serial Number for CompX Rack; repeat for each rack in compute-rack-definitions |
8888
| COMPX_RACK_LOCATION | Rack physical location for CompX Rack; repeat for each rack in compute-rack-definitions |
89-
| COMPX_SVRY_BMC_PASS | CompX Rack ServerY BMC password; repeat for each rack in compute-rack-definitions and for each server in rack |
89+
| COMPX_SVRY_BMC_PASS | CompX Rack ServerY Baseboard Management Controller (BMC) password; repeat for each rack in compute-rack-definitions and for each server in rack |
9090
| COMPX_SVRY_BMC_USER | CompX Rack ServerY BMC user; repeat for each rack in compute-rack-definitions and for each server in rack |
9191
| COMPX_SVRY_BMC_MAC | CompX Rack ServerY BMC MAC address; repeat for each rack in compute-rack-definitions and for each server in rack |
92-
| COMPX_SVRY_BOOT_MAC | CompX Rack ServerY boot NIC MAC address; repeat for each rack in compute-rack-definitions and for each server in rack |
92+
| COMPX_SVRY_BOOT_MAC | CompX Rack ServerY boot Network Interface Card (NIC) MAC address; repeat for each rack in compute-rack-definitions and for each server in rack |
9393
| COMPX_SVRY_SERVER_DETAILS | CompX Rack ServerY details; repeat for each rack in compute-rack-definitions and for each server in rack |
94-
| COMPX_SVRY_SERVER_NAME | CompX Rack ServerY name, repeat for each rack in compute-rack-definitions and for each server in rack |
94+
| COMPX_SVRY_SERVER_NAME | CompX Rack ServerY name; repeat for each rack in compute-rack-definitions and for each server in rack |
9595
| MRG_NAME | Cluster managed resource group name |
9696
| MRG_LOCATION | Cluster Azure region |
9797
| NF_ID | Reference to Network Fabric |
@@ -101,8 +101,8 @@ az networkcloud cluster create --name "$CLUSTER_NAME" --location "$LOCATION" \
101101
| TENANT_ID | Subscription tenant ID |
102102
| SUBSCRIPTION_ID | Subscription ID |
103103
| KV_RESOURCE_ID | Key Vault ID |
104-
| CLUSTER_TYPE | Type of cluster, Single, or MultiRack |
105-
| CLUSTER_VERSION | NC Version of cluster |
104+
| CLUSTER_TYPE | Type of cluster, Single, or MultiRack |
105+
| CLUSTER_VERSION | Network Cloud (NC) Version of cluster |
106106
| TAG_KEY1 | Optional tag1 to pass to Cluster Create |
107107
| TAG_VALUE1 | Optional tag1 value to pass to Cluster Create |
108108
| TAG_KEY2 | Optional tag2 to pass to Cluster Create |
@@ -131,23 +131,23 @@ You can find examples for an 8-Rack 2M16C SKU cluster using these two files:
131131
>[!NOTE]
132132
>To get the correct formatting, copy the raw code file. The values within the cluster.parameters.jsonc file are customer specific and may not be a complete list. Update the value fields for your specific environment.
133133
134-
1. In a web browser, go to the [Azure portal](https://portal.azure.com/) and sign in.
135-
1. From the Azure portal search bar, search for 'Deploy a custom template' and then select it from the available services.
134+
1. Navigate to [Azure portal](https://portal.azure.com/) in a web browser and sign in.
135+
1. Search for 'Deploy a custom template' in the Azure portal search bar, and then select it from the available services.
136136
1. Click on Build your own template in the editor.
137137
1. Click on Load file. Locate your cluster.jsonc template file and upload it.
138138
1. Click Save.
139139
1. Click Edit parameters.
140140
1. Click Load file. Locate your cluster.parameters.jsonc parameters file and upload it.
141141
1. Click Save.
142142
1. Select the correct Subscription.
143-
1. Search for the Resource group to see if it already exists. If not, create a new Resource group.
143+
1. Search for the Resource group to see if it already exists. If not, create a new Resource group.
144144
1. Make sure all Instance Details are correct.
145145
1. Click Review + create.
146146

147147

148148
### Cluster validation
149149

150-
A successful Operator Nexus Cluster creation results in the creation of an AKS cluster
150+
A successful Operator Nexus Cluster creation results in the creation of an Azure Kubernetes Service (AKS) cluster
151151
inside your subscription. The cluster ID, cluster provisioning state, and
152152
deployment state are returned as a result of a successful `cluster create`.
153153

@@ -170,16 +170,16 @@ Cluster create Logs can be viewed in the following locations:
170170

171171
## Deploy Cluster
172172

173-
After creating the cluster, the deploy cluster action can be triggered.
173+
The deploy Cluster action can be triggered after creating the Cluster.
174174
The deploy Cluster action creates the bootstrap image and deploys the Cluster.
175175

176176
Deploy Cluster initiates a sequence of events that occur in the Cluster Manager.
177177

178-
1. Validation of the cluster/rack properties
178+
1. Validation of the cluster/rack properties.
179179
2. Generation of a bootable image for the ephemeral bootstrap cluster
180180
(Validation of Infrastructure).
181-
3. Interaction with the IPMI interface of the targeted bootstrap machine.
182-
4. Perform hardware validation checks
181+
3. Interaction with the Intelligent Platform Management Interface (IPMI) interface of the targeted bootstrap machine.
182+
4. Performing hardware validation checks.
183183
5. Monitoring of the Cluster deployment process.
184184

185185
Deploy the on-premises Cluster:
@@ -198,7 +198,7 @@ az networkcloud cluster deploy \
198198
> See the section [Cluster Deploy Failed](#cluster-deploy-failed) for more detailed steps.
199199
> Optionally, the command can run asynchronously using the `--no-wait` flag.
200200
201-
### Cluster Deploy with hardware validation
201+
### Cluster Deployment with hardware validation
202202

203203
During a Cluster deploy process, one of the steps executed is hardware validation.
204204
The hardware validation procedure runs various test and checks against the machines
@@ -211,6 +211,9 @@ passed and/or are available to meet the thresholds necessary for deployment to c
211211
> Additionally, the provided Service Principal in the Cluster object is used for authentication against the Log Analytics Workspace Data Collection API.
212212
> This capability is only visible during a new deployment (Green Field); existing cluster will not have the logs available retroactively.
213213
214+
> [!NOTE]
215+
> The RAID controller is reset during Cluster deployment wiping all data from the server's virtual disks. Any Baseboard Management Controller (BMC) virtual disk alerts can typically be ignored unless there are additional physical disk and/or RAID controllers alerts.
216+
214217
By default, the hardware validation process writes the results to the configured Cluster `analyticsWorkspaceId`.
215218
However, due to the nature of Log Analytics Workspace data collection and schema evaluation, there can be ingestion delay that can take several minutes or more.
216219
For this reason, the Cluster deployment proceeds even if there was a failure to write the results to the Log Analytics Workspace.
@@ -219,9 +222,9 @@ To help address this possible event, the results, for redundancy, are also logge
219222
In the provided Cluster object's Log Analytics Workspace, a new custom table with the Cluster's name as prefix and the suffix `*_CL` should appear.
220223
In the _Logs_ section of the LAW resource, a query can be executed against the new `*_CL` Custom Log table.
221224

222-
#### Cluster Deploy Action with skipping specific bare-metal-machine
225+
#### Cluster Deployment with skipping specific bare-metal-machine
223226

224-
A parameter can be passed in to the deploy command that represents the names of
227+
The `--skip-validation-for-machines` parameter represents the names of
225228
bare metal machines in the cluster that should be skipped during hardware validation.
226229
Nodes skipped aren't validated and aren't added to the node pool.
227230
Additionally, nodes skipped don't count against the total used by threshold calculations.
@@ -279,7 +282,7 @@ az networkcloud cluster show --resource-group "$CLUSTER_RG" \
279282
```
280283

281284
The Cluster deployment is in-progress when detailedStatus is set to `Deploying` and detailedStatusMessage shows the progress of deployment.
282-
Some examples of deployment progress shown in detailedStatusMessage are `Hardware validation is in progress.` (if cluster is deployed with hardware validation) ,`Cluster is bootstrapping.`, `KCP initialization in progress.`, `Management plane deployment in progress.`, `Cluster extension deployment in progress.`, `waiting for "<rack-ids>" to be ready`, etc.
285+
Some examples of deployment progress shown in detailedStatusMessage are `Hardware validation is in progress.` (if cluster is deployed with hardware validation), `Cluster is bootstrapping.`, `KCP initialization in progress.`, `Management plane deployment in progress.`, `Cluster extension deployment in progress.`, `waiting for "<rack-ids>" to be ready`, etc.
283286

284287
:::image type="content" source="./media/nexus-deploy-kcp-status.png" lightbox="./media/nexus-deploy-kcp-status.png" alt-text="Screenshot of Azure portal showing cluster deploy progress kcp init.":::
285288

@@ -372,7 +375,7 @@ Note, `<APIVersion>` is the API version 2024-07-01 or newer.
372375

373376
## Delete a cluster
374377

375-
When deleting a cluster, it deletes the resources in Azure and the cluster that resides in the on-premises environment.
378+
Deleting a cluster deletes the resources in Azure and the cluster that resides in the on-premises environment.
376379

377380
>[!NOTE]
378381
>If there are any tenant resources that exist in the cluster, it will not be deleted until those resources are deleted.

0 commit comments

Comments
 (0)