diff --git a/modules/authentication/proc-creating-a-custom-transformer-to-provision-users-from-rhbk-to-the-software-catalog.adoc b/modules/authentication/proc-creating-a-custom-transformer-to-provision-users-from-rhbk-to-the-software-catalog.adoc index 3e2ea37164..7aa0ddde09 100644 --- a/modules/authentication/proc-creating-a-custom-transformer-to-provision-users-from-rhbk-to-the-software-catalog.adoc +++ b/modules/authentication/proc-creating-a-custom-transformer-to-provision-users-from-rhbk-to-the-software-catalog.adoc @@ -70,7 +70,7 @@ Set the module's `pluginId` to `catalog` to match the `pluginId` of the `keycloa + [source,javascript] ---- -backend.add(import(backstage-plugin-catalog-backend-module-keycloak-transformer)) +$ backend.add(import(backstage-plugin-catalog-backend-module-keycloak-transformer)) ---- .Verification diff --git a/modules/observe/adoption-insights/proc-customize-adoption-insights.adoc b/modules/observe/adoption-insights/proc-customize-adoption-insights.adoc index d612df5817..caaebf078a 100644 --- a/modules/observe/adoption-insights/proc-customize-adoption-insights.adoc +++ b/modules/observe/adoption-insights/proc-customize-adoption-insights.adoc @@ -12,7 +12,7 @@ You can customize the Adoption Insights plugin to suit your needs by disabling o * To customize `maxBufferSize`, `flushInterval`, `debug`, and `licensedUsers` in the Adoption Insights plugin, in your {product} `app-config.yaml` file, update the relevant settings as shown in the following code: + -[source,terminal] +[source,yaml] ---- app: analytics: diff --git a/modules/observe/proc-admin-enabling-metrics-ocp-operator.adoc b/modules/observe/proc-admin-enabling-metrics-ocp-operator.adoc index 3878ff0be1..112c310025 100644 --- a/modules/observe/proc-admin-enabling-metrics-ocp-operator.adoc +++ b/modules/observe/proc-admin-enabling-metrics-ocp-operator.adoc @@ -19,7 +19,7 @@ By setting the `spec.monitoring.enabled` field to `true` in your {product} custo + [source,bash] ---- -oc edit Backstage +$ oc edit Backstage ---- . In the CR, locate the `spec` field and add the `monitoring` configuration block. + diff --git a/modules/observe/proc-configure-annotations-for-aks-monitoring.adoc b/modules/observe/proc-configure-annotations-for-aks-monitoring.adoc index ff82c0ab4f..77f2fddc69 100644 --- a/modules/observe/proc-configure-annotations-for-aks-monitoring.adoc +++ b/modules/observe/proc-configure-annotations-for-aks-monitoring.adoc @@ -34,7 +34,7 @@ Operator-backed deployment:: ---- # Update OPERATOR_NS accordingly OPERATOR_NS=rhdh-operator -kubectl edit configmap backstage-default-config -n "${OPERATOR_NS}" +$ kubectl edit configmap backstage-default-config -n "${OPERATOR_NS}" ---- . Find the `deployment.yaml` key in the ConfigMap and add the annotations to the `spec.template.metadata.annotations` field as follows: diff --git a/modules/observe/proc-configuring-annotations-for-monitoring-with-amazon-prometheus-by-using-the-helm-chart.adoc b/modules/observe/proc-configuring-annotations-for-monitoring-with-amazon-prometheus-by-using-the-helm-chart.adoc index 821253a8b9..f10cca52dd 100644 --- a/modules/observe/proc-configuring-annotations-for-monitoring-with-amazon-prometheus-by-using-the-helm-chart.adoc +++ b/modules/observe/proc-configuring-annotations-for-monitoring-with-amazon-prometheus-by-using-the-helm-chart.adoc @@ -27,7 +27,7 @@ To verify if the scraping works: + [source,bash] ---- -kubectl --namespace=prometheus port-forward deploy/prometheus-server 9090 +$ kubectl --namespace=prometheus port-forward deploy/prometheus-server 9090 ---- . Open your web browser and navigate to `pass:c[http://localhost:9090]` to access the Prometheus console. diff --git a/modules/observe/proc-enabling-azure-monitor-metrics.adoc b/modules/observe/proc-enabling-azure-monitor-metrics.adoc index 44572b63e6..0151f680d3 100644 --- a/modules/observe/proc-enabling-azure-monitor-metrics.adoc +++ b/modules/observe/proc-enabling-azure-monitor-metrics.adoc @@ -7,7 +7,7 @@ To enable managed Prometheus monitoring, use the `-enable-azure-monitor-metrics` [source,bash] ---- -az aks create/update --resource-group --name --enable-azure-monitor-metrics +$ az aks create/update --resource-group --name --enable-azure-monitor-metrics ---- The previous command installs the metrics add-on, which gathers https://learn.microsoft.com/en-us/azure/azure-monitor/metrics/prometheus-metrics-overview[Prometheus metrics]. Using the previous command, you can enable monitoring of Azure resources through both native Azure Monitor metrics. You can also view the results in the portal under *Monitoring -> Insights*. For more information, see https://learn.microsoft.com/en-us/azure/azure-monitor/platform/monitor-azure-resource[Monitor Azure resources with Azure Monitor]. diff --git a/modules/observe/proc-forward-audit-log-splunk.adoc b/modules/observe/proc-forward-audit-log-splunk.adoc index 571978e059..ee9bea3b43 100644 --- a/modules/observe/proc-forward-audit-log-splunk.adoc +++ b/modules/observe/proc-forward-audit-log-splunk.adoc @@ -20,7 +20,7 @@ You can use the {logging-brand-name} ({logging-short}) Operator and a `ClusterLo .Example command to switch to a namespace [source,bash] ---- -oc project openshift-logging +$ oc project openshift-logging ---- -- . Create a `serviceAccount` named `log-collector` and bind the `collect-application-logs` role to the `serviceAccount` : @@ -29,13 +29,13 @@ oc project openshift-logging .Example command to create a `serviceAccount` [source,bash] ---- -oc create sa log-collector +$ oc create sa log-collector ---- .Example command to bind a role to a `serviceAccount` [source,bash] ---- -oc create clusterrolebinding log-collector --clusterrole=collect-application-logs --serviceaccount=openshift-logging:log-collector +$ oc create clusterrolebinding log-collector --clusterrole=collect-application-logs --serviceaccount=openshift-logging:log-collector ---- -- . Generate a `hecToken` in your Splunk instance. @@ -45,13 +45,13 @@ oc create clusterrolebinding log-collector --clusterrole=collect-application-log .Example command to create a key/value secret with `hecToken` [source,bash] ---- -oc -n openshift-logging create secret generic splunk-secret --from-literal=hecToken= +$ oc -n openshift-logging create secret generic splunk-secret --from-literal=hecToken= ---- .Example command to verify a secret [source,bash] ---- -oc -n openshift-logging get secret/splunk-secret -o yaml +$ oc -n openshift-logging get secret/splunk-secret -o yaml ---- -- . Create a basic `ClusterLogForwarder`resource YAML file as follows: @@ -160,7 +160,7 @@ pipelines: .Example command to apply `ClusterLogForwarder` configuration [source,bash] ---- -oc apply -f +$ oc apply -f ---- -- . Optional: To reduce the risk of log loss, configure your `ClusterLogForwarder` pods using the following options: diff --git a/modules/observe/scorecards/proc-configuring-jira-scorecards-in-rhdh-instance.adoc b/modules/observe/scorecards/proc-configuring-jira-scorecards-in-rhdh-instance.adoc index 409980d286..b3bd8c03ee 100644 --- a/modules/observe/scorecards/proc-configuring-jira-scorecards-in-rhdh-instance.adoc +++ b/modules/observe/scorecards/proc-configuring-jira-scorecards-in-rhdh-instance.adoc @@ -30,7 +30,7 @@ You must use the proxy setup to ensure configuration compatibility if the Roadie + [source,bash] ---- -echo -n 'your-atlassian-email:your-jira-api-token' | base64 +$ echo -n 'your-atlassian-email:your-jira-api-token' | base64 ---- ** Jira datacenter: Create and use a Personal Access Token (PAT) in your Jira datacenter account. For more information, see the https://confluence.atlassian.com/enterprise/using-personal-access-tokens-1026032365.html[Atlassian] documentation. diff --git a/modules/openshift-ai-connector-for-rhdh/proc-populating-the-api-definition-tab.adoc b/modules/openshift-ai-connector-for-rhdh/proc-populating-the-api-definition-tab.adoc index b71596cde7..34ec40e353 100644 --- a/modules/openshift-ai-connector-for-rhdh/proc-populating-the-api-definition-tab.adoc +++ b/modules/openshift-ai-connector-for-rhdh/proc-populating-the-api-definition-tab.adoc @@ -3,7 +3,7 @@ [id="proc-populating-the-api-definition-tab_{context}"] = Populating the API Definition tab in {product-very-short} API entities -Since {rhoai-short} does not expose the OpenAPI specification by default, the AI platform engineer can take the following steps to provide this valuable information: +Because {rhoai-short} does not expose the OpenAPI specification by default, the AI platform engineer can take the following steps to provide this valuable information: .Procedure @@ -11,7 +11,7 @@ Since {rhoai-short} does not expose the OpenAPI specification by default, the AI + [source,bash] ---- -curl -k -H "Authorization: Bearer $MODEL_API_KEY" https://$MODEL_ROOT_URL_INCLUDING_PORT/openapi.json | jq > open-api.json +$ curl -k -H "Authorization: Bearer $MODEL_API_KEY" https://$MODEL_ROOT_URL_INCLUDING_PORT/openapi.json | jq > open-api.json ---- . Set Property in {rhoai-short}. @@ -24,4 +24,4 @@ We recommend using *Model Version* instead of *Registered Model* to maintain sta .. In the **Properties** section, set a key/value pair where the key is `API Spec` and the value is the entire JSON content from the `open-api.json` file. -. Propagation: The {openshift-ai-connector-name} periodically polls the {rhoai-short} Model Registry, propagates this JSON, and renders the interactive API documentation in the *Definition* tab of the {product-very-short} API entity. \ No newline at end of file +. Propagation: The {openshift-ai-connector-name} periodically polls the {rhoai-short} Model Registry, propagates this JSON, and renders the interactive API documentation in the *Definition* tab of the {product-very-short} API entity. diff --git a/modules/openshift-ai-connector-for-rhdh/proc-troubleshooting-connector-functionality.adoc b/modules/openshift-ai-connector-for-rhdh/proc-troubleshooting-connector-functionality.adoc index 858d10f742..6ad69bee18 100644 --- a/modules/openshift-ai-connector-for-rhdh/proc-troubleshooting-connector-functionality.adoc +++ b/modules/openshift-ai-connector-for-rhdh/proc-troubleshooting-connector-functionality.adoc @@ -18,7 +18,7 @@ Validate that the dynamic plugins have been successfully installed into your {pr [source,bash,subs=+attributes] ---- -oc logs -c install-dynamic-plugins deployment/ +$ oc logs -c install-dynamic-plugins deployment/ ---- The `install-dynamic-plugin` logs allow you to check the following installation logs for successful logs: @@ -53,22 +53,22 @@ The {openshift-ai-connector-name-short} sidecars manage the data fetching and st + [source,bash] ---- -oc get configmap bac-import-model -o json | jq -r '.binaryData | to_entries[] | "=== \(.key) ===\n" + (.value | @base64d | fromjson | .body | @base64d | fromjson | tostring)' | jq -R 'if startswith("=== ") then . else (. | fromjson) end' +$ oc get configmap bac-import-model -o json | jq -r '.binaryData | to_entries[] | "=== \(.key) ===\n" + (.value | @base64d | fromjson | .body | @base64d | fromjson | tostring)' | jq -R 'if startswith("=== ") then . else (. | fromjson) end' ---- . Check Location Service API: Confirm the location service is providing data to the {product-very-short} Entity Provider. + [source,bash,subs=+attributes] ---- -oc rsh -c backstage-backend deployment/ -curl http://localhost:9090/list +$ oc rsh -c backstage-backend deployment/ +$ curl http://localhost:9090/list ---- . Check Sidecar Container Logs: + [source,bash] ---- -oc logs -c rhoai-normalizer deployment/ -oc logs -c storage-rest deployment/ -oc logs -c location deployment/ +$ oc logs -c rhoai-normalizer deployment/ +$ oc logs -c storage-rest deployment/ +$ oc logs -c location deployment/ ---- \ No newline at end of file diff --git a/modules/openshift-ai-connector-for-rhdh/ref-openshift-ai-model-registry-and-model-catalog-queries.adoc b/modules/openshift-ai-connector-for-rhdh/ref-openshift-ai-model-registry-and-model-catalog-queries.adoc index b87b896b2b..437a9eebd2 100644 --- a/modules/openshift-ai-connector-for-rhdh/ref-openshift-ai-model-registry-and-model-catalog-queries.adoc +++ b/modules/openshift-ai-connector-for-rhdh/ref-openshift-ai-model-registry-and-model-catalog-queries.adoc @@ -9,40 +9,40 @@ To access the same {rhoai-short} data as the connector, use `curl` to query the + [source,bash] ---- -curl -k -H "Authorization: Bearer $TOKEN" $RHOAI_MODEL_REGISTRY_URL/api/model_registry/v1alpha3/registered_models | jq +$ curl -k -H "Authorization: Bearer $TOKEN" $RHOAI_MODEL_REGISTRY_URL/api/model_registry/v1alpha3/registered_models | jq ---- * Example showing how to fetch model versions + [source,bash] ---- -curl -k -H "Authorization: Bearer $TOKEN" $RHOAI_MODEL_REGISTRY_URL/api/model_registry/v1alpha3/model_versions | jq +$ curl -k -H "Authorization: Bearer $TOKEN" $RHOAI_MODEL_REGISTRY_URL/api/model_registry/v1alpha3/model_versions | jq ---- * Example showing how to fetch model artifacts + [source,bash] ---- -curl -k -H "Authorization: Bearer $TOKEN" $RHOAI_MODEL_REGISTRY_URL/api/model_registry/v1alpha3/model_artifacts | jq +$ curl -k -H "Authorization: Bearer $TOKEN" $RHOAI_MODEL_REGISTRY_URL/api/model_registry/v1alpha3/model_artifacts | jq ---- * Example showing how to fetch inference services + [source,bash] ---- -curl -k -H "Authorization: Bearer $TOKEN" $RHOAI_MODEL_REGISTRY_URL/api/model_registry/v1alpha3/inference_services | jq +$ curl -k -H "Authorization: Bearer $TOKEN" $RHOAI_MODEL_REGISTRY_URL/api/model_registry/v1alpha3/inference_services | jq ---- * Example showing how to fetch serving environments + [source,bash] ---- -curl -k -H "Authorization: Bearer $TOKEN" $RHOAI_MODEL_REGISTRY_URL/api/model_registry/v1alpha3/serving_environments | jq +$ curl -k -H "Authorization: Bearer $TOKEN" $RHOAI_MODEL_REGISTRY_URL/api/model_registry/v1alpha3/serving_environments | jq ---- * Example showing how to fetch catalog sources + [source,bash] ---- -curl -k -H "Authorization: Bearer $TOKEN" $RHOAI_MODEL_CATALOG_URL/api/model_catalog/v1alpha1/sources | jq +$ curl -k -H "Authorization: Bearer $TOKEN" $RHOAI_MODEL_CATALOG_URL/api/model_catalog/v1alpha1/sources | jq ---- \ No newline at end of file diff --git a/modules/orchestrator/proc-building-the-01-basic-workflow.adoc b/modules/orchestrator/proc-building-the-01-basic-workflow.adoc index 08e9f05f5e..e3b30c0d88 100644 --- a/modules/orchestrator/proc-building-the-01-basic-workflow.adoc +++ b/modules/orchestrator/proc-building-the-01-basic-workflow.adoc @@ -15,7 +15,7 @@ To run the script from the root directory of the repository, you must use the `- + [source,bash] ---- -./scripts/build.sh --image=quay.io/orchestrator/demo-basic:test -w 01_basic/ -m 01_basic/manifests +$ ./scripts/build.sh --image=quay.io/orchestrator/demo-basic:test -w 01_basic/ -m 01_basic/manifests ---- + This build command produces the following two artifacts: diff --git a/modules/orchestrator/proc-creating-and-running-workflows.adoc b/modules/orchestrator/proc-creating-and-running-workflows.adoc index ae11544098..a0242306dd 100644 --- a/modules/orchestrator/proc-creating-and-running-workflows.adoc +++ b/modules/orchestrator/proc-creating-and-running-workflows.adoc @@ -10,14 +10,14 @@ The `kn-workflow` CLI is an essential tool that generates workflow manifests and + [source,bash] ---- -kn-workflow quarkus create --name +$ kn-workflow quarkus create --name ---- . Edit the workflow, add schema and specific files, and run it locally from project folder as shown in the following example: + [source,bash] ---- -kn-workflow quarkus run +$ kn-workflow quarkus run ---- . Run the workflow locally using the `kn-workflow run` which pulls the following image: + diff --git a/modules/orchestrator/proc-deploying-workflows-on-a-cluster.adoc b/modules/orchestrator/proc-deploying-workflows-on-a-cluster.adoc index fbc4c5df92..7bdf8dbde1 100644 --- a/modules/orchestrator/proc-deploying-workflows-on-a-cluster.adoc +++ b/modules/orchestrator/proc-deploying-workflows-on-a-cluster.adoc @@ -3,7 +3,7 @@ [id="proc-deploying-workflows-on-a-cluster_{context}"] = Deploying workflows on a cluster -You can deploy the workflow on a cluster, since the image is pushed to the image registry and the deployment manifests are available. +You can deploy the workflow on a cluster, because the image is pushed to the image registry and the deployment manifests are available. .Prerequisites @@ -24,23 +24,23 @@ For instructions on how to install these components, see the {orchestrator-book- + [source,bash] ---- -kubectl create -n -f ./01_basic/manifests/. +$ kubectl create -n -f ./01_basic/manifests/. ---- . After deployment, monitor the status of the workflow pods as shown in the following example: + -[source,yaml] +[source,bash] ---- -kubectl get pods -n -l app=basic +$ kubectl get pods -n -l app=basic ---- + The pod may initially appear in an `Error` state because of missing or incomplete configuration in the Secret or ConfigMap. . Inspect the Pod logs as shown in the following example: + -[source,yaml] +[source,bash] ---- -oc logs -n basic-f7c6ff455-vwl56 +$ oc logs -n basic-f7c6ff455-vwl56 ---- + The following code is an example of the output: @@ -57,9 +57,9 @@ The error indicates a missing property: `quarkus.openapi-generator.notifications . In such a case where the logs show the `ConfigurationException: Failed to read configuration properties` error or indicate a missing value, retrieve the ConfigMap as shown in the following example: + -[source,yaml] +[source,bash] ---- -oc get -n configmaps basic-props -o yaml +$ oc get -n configmaps basic-props -o yaml ---- + The following code is an example of the sample output: @@ -79,9 +79,9 @@ Resolve the placeholders using values provided using a Secret. . You must edit the corresponding Secret and provide appropriate base64-encoded values to resolve the placeholders in `application.properties` as shown in the following example: + -[source,yaml] +[source,bash] ---- -kubectl edit secrets -n basic-secrets +$ kubectl edit secrets -n basic-secrets ---- . Restart the workflow Pod for Secret changes to take effect in OpenShift Serverless Logic `v1.36`. @@ -89,9 +89,9 @@ kubectl edit secrets -n basic-secrets . Verify the deployment status by checking the Pods again as shown in the following example: + -[source,yaml] +[source,bash] ---- -oc get pods -n -l app=basic +$ oc get pods -n -l app=basic ---- + The expected status for a successfully deployed workflow Pod is as shown in the following example: diff --git a/modules/orchestrator/proc-enable-orchestrator-plugin.adoc b/modules/orchestrator/proc-enable-orchestrator-plugin.adoc index e66dc9b093..026d150b64 100644 --- a/modules/orchestrator/proc-enable-orchestrator-plugin.adoc +++ b/modules/orchestrator/proc-enable-orchestrator-plugin.adoc @@ -26,7 +26,7 @@ Additionally, the `ref: sonataflow` field installs the Openshift Serverless and ==== + .Example: Complete configuration of the Orchestrator plugin -[source,subs="+attributes,+quotes"] +[source,yaml,subs="+attributes,+quotes"] ---- apiVersion: v1 kind: ConfigMap diff --git a/modules/orchestrator/proc-helm-install-components-orchestrator-plugin.adoc b/modules/orchestrator/proc-helm-install-components-orchestrator-plugin.adoc index a6cb091f8d..44767872ae 100644 --- a/modules/orchestrator/proc-helm-install-components-orchestrator-plugin.adoc +++ b/modules/orchestrator/proc-helm-install-components-orchestrator-plugin.adoc @@ -11,7 +11,7 @@ You can use Orchestrator Infrastructure for {product} to install components for + [source,terminal,subs="attributes+"] ---- -helm install redhat-developer/redhat-developer-hub-orchestrator-infra +$ helm install redhat-developer/redhat-developer-hub-orchestrator-infra ---- + [NOTE] diff --git a/modules/orchestrator/proc-helper-script-overview.adoc b/modules/orchestrator/proc-helper-script-overview.adoc index 824535b122..e8fd39be33 100644 --- a/modules/orchestrator/proc-helper-script-overview.adoc +++ b/modules/orchestrator/proc-helper-script-overview.adoc @@ -16,7 +16,7 @@ Do not use `plugin-infra.sh` in production. + [source,terminal,subs="+attributes,+quotes"] ---- -curl -sSLO https://raw.githubusercontent.com/redhat-developer/rhdh-operator/refs/heads/release-{product-version}/config/profile/rhdh/plugin-infra/plugin-infra.sh # Specify the {product} version in the URL or use main +$ curl -sSLO https://raw.githubusercontent.com/redhat-developer/rhdh-operator/refs/heads/release-{product-version}/config/profile/rhdh/plugin-infra/plugin-infra.sh # Specify the {product} version in the URL or use main ---- . Run the script: diff --git a/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-helm-chart-full.adoc b/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-helm-chart-full.adoc index 27b68be461..20b8db0b57 100644 --- a/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-helm-chart-full.adoc +++ b/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-helm-chart-full.adoc @@ -70,7 +70,7 @@ yq -r '.data."controllers_cfg.yaml" | from_yaml | .. | select(tag == "!!str") | + [source,terminal,subs="+attributes,+quotes"] ---- -oc-mirror --config=ImageSetConfiguration.yaml file:///path/to/mirror-archive --authfile /path/to/authfile --v2 +$ oc-mirror --config=ImageSetConfiguration.yaml file:///path/to/mirror-archive --authfile /path/to/authfile --v2 ---- + [NOTE] @@ -82,8 +82,8 @@ The `oc-mirror` command pulls the charts listed in the `ImageSetConfiguration` f + [source,terminal,subs="+attributes,+quotes"] ---- -cd /working-dir/cluster-resources/ -oc apply -f . +$ cd /working-dir/cluster-resources/ +$ oc apply -f . ---- + . Transfer the generated mirror archive file, for example, `/path/to/mirror-archive/mirror_000001.tar`, to a bastion host within your disconnected environment. @@ -92,7 +92,7 @@ oc apply -f . + [source,terminal,subs="+attributes,+quotes"] ---- -oc-mirror --v2 --from docker:// --workspace file:// --authfile /path/to/authfile +$ oc-mirror --v2 --from docker:// --workspace file:// --authfile /path/to/authfile ---- + where: diff --git a/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-helm-chart-partial.adoc b/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-helm-chart-partial.adoc index fb5dd586cd..bbcf3b138d 100644 --- a/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-helm-chart-partial.adoc +++ b/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-helm-chart-partial.adoc @@ -60,7 +60,7 @@ mirror: + [source,terminal,subs="+attributes,+quotes"] ---- -oc-mirror --config=imagesetconfiguration.yaml docker:// --workspace file:// --authfile /path/to/authfile --v2 +$ oc-mirror --config=imagesetconfiguration.yaml docker:// --workspace file:// --authfile /path/to/authfile --v2 ---- + [NOTE] @@ -72,8 +72,8 @@ The `oc-mirror` command pulls the charts listed in the `ImageSetConfiguration` f + [source,terminal,subs="+attributes,+quotes"] ---- -cd /working-dir/cluster-resources/ -oc apply -f . +$ cd /working-dir/cluster-resources/ +$ oc apply -f . ---- + include::snip-installing-the-orchestrator-common-steps.adoc[] diff --git a/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-operator-full.adoc b/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-operator-full.adoc index 94778e1cf1..10f05937da 100644 --- a/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-operator-full.adoc +++ b/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-operator-full.adoc @@ -64,7 +64,7 @@ yq -r '.data."controllers_cfg.yaml" | from_yaml | .. | select(tag == "!!str") | + [source,terminal,subs="+attributes,+quotes"] ---- -oc-mirror --config=ImageSetConfiguration.yaml file:///path/to/mirror-archive --authfile /path/to/authfile --v2 +$ oc-mirror --config=ImageSetConfiguration.yaml file:///path/to/mirror-archive --authfile /path/to/authfile --v2 ---- + [NOTE] @@ -78,7 +78,7 @@ The `oc-mirror` command generates a local workspace containing the mirror archiv + [source,yaml,subs="+attributes,+quotes"] ---- -oc-mirror --v2 --from docker:// --workspace file:// --authfile /path/to/authfile +$ oc-mirror --v2 --from docker:// --workspace file:// --authfile /path/to/authfile ---- + where: @@ -91,8 +91,8 @@ where: + [source,terminal,subs="+attributes,+quotes"] ---- -cd /working-dir/cluster-resources/ -oc apply -f . +$ cd /working-dir/cluster-resources/ +$ oc apply -f . ---- include::snip-installing-the-orchestrator-common-steps.adoc[] diff --git a/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-operator-partial.adoc b/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-operator-partial.adoc index fd28876b85..a9014ba2c3 100644 --- a/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-operator-partial.adoc +++ b/modules/orchestrator/proc-install-rhdh-orchestrator-airgapped-env-using-operator-partial.adoc @@ -64,9 +64,9 @@ yq -r '.data."controllers_cfg.yaml" | from_yaml | .. | select(tag == "!!str") | + [source,terminal,subs="+attributes,+quotes"] ---- -oc-mirror --config=imagesetconfiguration.yaml docker:// --workspace file:// --authfile /path/to/authfile --v2 -cd /working-dir/cluster-resources/ -oc apply -f . +$ oc-mirror --config=imagesetconfiguration.yaml docker:// --workspace file:// --authfile /path/to/authfile --v2 +$ cd /working-dir/cluster-resources/ +$ oc apply -f . ---- + include::snip-installing-the-orchestrator-common-steps.adoc[] diff --git a/modules/orchestrator/proc-install-rhdh-with-orchestrator-helm-cli.adoc b/modules/orchestrator/proc-install-rhdh-with-orchestrator-helm-cli.adoc index d5b19055c5..bd3209ac4a 100644 --- a/modules/orchestrator/proc-install-rhdh-with-orchestrator-helm-cli.adoc +++ b/modules/orchestrator/proc-install-rhdh-with-orchestrator-helm-cli.adoc @@ -23,10 +23,10 @@ By default, Orchestrator Infrastructure for {product} Helm chart does *not* auto . As an administrator, install relevant cluster-wide resources. + -[source,yaml,subs="+quotes,+attributes"] +[source,terminal,subs="+quotes,+attributes"] ---- -helm repo add openshift-helm-charts https://charts.openshift.io/ -helm install `` openshift-helm-charts/redhat-developer-hub-orchestrator-infra +$ helm repo add openshift-helm-charts https://charts.openshift.io/ +$ helm install `` openshift-helm-charts/redhat-developer-hub-orchestrator-infra ---- + [IMPORTANT] @@ -63,7 +63,7 @@ global: + [source,terminal,subs="attributes+"] ---- -helm install openshift-helm-charts/redhat-developer-hub \ +$ helm install openshift-helm-charts/redhat-developer-hub \ --version {product-chart-version} \ --set orchestrator.enabled=true \ --set orchestrator.serverlessOperator=false \ diff --git a/modules/orchestrator/proc-upgrading-the-orchestrator-plugin.adoc b/modules/orchestrator/proc-upgrading-the-orchestrator-plugin.adoc index 5c4da3e325..9799bbafc7 100644 --- a/modules/orchestrator/proc-upgrading-the-orchestrator-plugin.adoc +++ b/modules/orchestrator/proc-upgrading-the-orchestrator-plugin.adoc @@ -51,7 +51,7 @@ The upgrade takes a few minutes to complete. The {product} version does not upda + [source,terminal] ---- -oc get pods -w +$ oc get pods -w ---- + The upgrade is successful when all {product-custom-resource-type}-related pods show a stable *Running* status. \ No newline at end of file diff --git a/modules/orchestrator/ref-orchestrator-resource-limits.adoc b/modules/orchestrator/ref-orchestrator-resource-limits.adoc index 2f98c54f7f..230d227a87 100644 --- a/modules/orchestrator/ref-orchestrator-resource-limits.adoc +++ b/modules/orchestrator/ref-orchestrator-resource-limits.adoc @@ -38,7 +38,7 @@ orchestrator: + [source,terminal] ---- -helm upgrade --install openshift-helm-charts/redhat-developer-hub \ +$ helm upgrade --install openshift-helm-charts/redhat-developer-hub \ --set orchestrator.enabled=true \ --set orchestrator.sonataflowPlatform.resources.requests.cpu=500m \ --set orchestrator.sonataflowPlatform.resources.requests.memory=128Mi \ diff --git a/modules/orchestrator/snip-installing-the-orchestrator-common-steps.adoc b/modules/orchestrator/snip-installing-the-orchestrator-common-steps.adoc index 34abbd1305..f33c42b491 100644 --- a/modules/orchestrator/snip-installing-the-orchestrator-common-steps.adoc +++ b/modules/orchestrator/snip-installing-the-orchestrator-common-steps.adoc @@ -9,18 +9,18 @@ + [source,subs="+attributes,+quotes"] ---- -npm pack "@redhat/backstage-plugin-orchestrator@{orchestrator-plugin-version}" --registry=https://npm.registry.redhat.com -npm pack "@redhat/backstage-plugin-orchestrator-backend-dynamic@{orchestrator-plugin-version}" --registry=https://npm.registry.redhat.com -npm pack "@redhat/backstage-plugin-scaffolder-backend-module-orchestrator-dynamic@{orchestrator-plugin-version} --registry=https://npm.registry.redhat.com -npm pack "@redhat/backstage-plugin-orchestrator-form-widgets@{orchestrator-plugin-version}" --registry=https://npm.registry.redhat.com +$ npm pack "@redhat/backstage-plugin-orchestrator@{orchestrator-plugin-version}" --registry=https://npm.registry.redhat.com +$ npm pack "@redhat/backstage-plugin-orchestrator-backend-dynamic@{orchestrator-plugin-version}" --registry=https://npm.registry.redhat.com +$ npm pack "@redhat/backstage-plugin-scaffolder-backend-module-orchestrator-dynamic@{orchestrator-plugin-version} --registry=https://npm.registry.redhat.com +$ npm pack "@redhat/backstage-plugin-orchestrator-form-widgets@{orchestrator-plugin-version}" --registry=https://npm.registry.redhat.com ---- . Push the NPM packages you have downloaded to your NPM server, as shown in the following example: + [source,subs="+attributes,+quotes"] ---- -npm publish backstage-plugin-orchestrator-{orchestrator-plugin-version}.tgz -npm publish backstage-plugin-orchestrator-backend-dynamic-{orchestrator-plugin-version}.tgz -npm publish backstage-plugin-orchestrator-form-widgets-{orchestrator-plugin-version}.tgz -npm publish backstage-plugin-scaffolder-backend-module-orchestrator-dynamic-{orchestrator-plugin-version}.tgz +$ npm publish backstage-plugin-orchestrator-{orchestrator-plugin-version}.tgz +$ npm publish backstage-plugin-orchestrator-backend-dynamic-{orchestrator-plugin-version}.tgz +$ npm publish backstage-plugin-orchestrator-form-widgets-{orchestrator-plugin-version}.tgz +$ npm publish backstage-plugin-scaffolder-backend-module-orchestrator-dynamic-{orchestrator-plugin-version}.tgz ---- \ No newline at end of file diff --git a/modules/proc-create-plugin.adoc b/modules/proc-create-plugin.adoc index c1515b4c3f..bd6cd83463 100644 --- a/modules/proc-create-plugin.adoc +++ b/modules/proc-create-plugin.adoc @@ -17,7 +17,7 @@ You can create a plugin for the {product-short} application. -- [source,terminal] ---- -yarn new --select plugin +$ yarn new --select plugin ---- The previous command creates a new plugin based on the provided ID. The plugin is added to the {product-short} application automatically. @@ -30,7 +30,7 @@ The previous command creates a new plugin based on the provided ID. The plugin i -- [source,bash] ---- -yarn workspace {pluginID} start # Also supports --check <1> +$ yarn workspace {pluginID} start # Also supports --check <1> ---- <1> ID of the new plugin such as `@red-hat-developer-hub/my-new-plugin` diff --git a/modules/techdocs/proc-techdocs-addon-create.adoc b/modules/techdocs/proc-techdocs-addon-create.adoc index ec8667bcff..d00d94a842 100644 --- a/modules/techdocs/proc-techdocs-addon-create.adoc +++ b/modules/techdocs/proc-techdocs-addon-create.adoc @@ -74,23 +74,23 @@ Upon completion of this action, a sub-directory with the same name that you gave + [source,terminal,subs="+attributes,+quotes"] ---- -cd plugins/__ +$ cd plugins/__ ---- . Add the`@backstage/plugin-techdocs-react` package to get frontend utilities for TechDocs add-ons. For example: + [source,terminal,subs="+attributes,+quotes"] ---- -yarn add @backstage/plugin-techdocs-react +$ yarn add @backstage/plugin-techdocs-react ---- . In the directory containing the components of your custom TechDocs add-on, delete any default files or file components that are not required for your add-on, such as the `routes.ts` file or components of the `index.tsx` and `plugins.ts` files. . In the `plugins.ts` file, add the following code: + [source,java,subs="+attributes,+quotes"] ---- -import { createPlugin } from '@backstage/core-plugin-api'; -import { createTechDocsAddonExtension } from '@backstage/plugin-techdocs-react'; +$ import { createPlugin } from '@backstage/core-plugin-api'; +$ import { createTechDocsAddonExtension } from '@backstage/plugin-techdocs-react'; -export const __ = createPlugin({ +$ export const __ = createPlugin({ id: '__', }); @@ -98,7 +98,7 @@ export const __ = createPlugin({ * * @public */ -export const __ = __.provide( +$ export const __ = __.provide( createTechDocsAddonExtension<__>({ name: '__', location: TechDocsAddonLocations.Content, @@ -117,7 +117,7 @@ __ :: Specifies the React component for the custo + [source,java,subs="+attributes,+quotes"] ---- -export { __, __ } from './plugin'; +$ export { __, __ } from './plugin'; ---- . Create a new `__.tsx` file and add the code for your new TechDocs add-on component. + @@ -132,7 +132,7 @@ can also mention a template that the user can configure, if there is one + [source,java,subs="+attributes,+quotes"] ---- -export { __, type __} from './__' +$ export { __, type __} from './__' ---- + where diff --git a/modules/techdocs/proc-techdocs-addon-install-third-party.adoc b/modules/techdocs/proc-techdocs-addon-install-third-party.adoc index 61307b7fd9..9d802ef0f9 100644 --- a/modules/techdocs/proc-techdocs-addon-install-third-party.adoc +++ b/modules/techdocs/proc-techdocs-addon-install-third-party.adoc @@ -16,14 +16,14 @@ You can install compatible third-party TechDocs add-on on your {product} instanc + [source,terminal,subs="+quotes,+attributes"] ---- -yarn install +$ yarn install ---- . Obtain the source code for the third-party TechDocs add-on that you want to use. . Export the TechDocs add-on as a dynamic plugin using the following command: + [source,terminal,subs="+quotes,+attributes"] ---- -npx @red-hat-developer-hub/cli@latest plugin export +$ npx @red-hat-developer-hub/cli@latest plugin export ---- + [NOTE] @@ -34,7 +34,7 @@ The `@latest` tag pulls the latest version of the @red-hat-developer-hub/cli pac + [source,terminal,subs="+quotes,+attributes"] ---- -npx @red-hat-developer-hub/cli@latest plugin package --tag quay.io/__/__:latest +$ npx @red-hat-developer-hub/cli@latest plugin package --tag quay.io/__/__:latest ---- + [NOTE] @@ -47,13 +47,13 @@ The output of the package-dynamic-plugins command provides the file path to the + [source,terminal,subs="+quotes,+attributes"] ---- -podman push quay.io/__/__:latest +$ podman push quay.io/__/__:latest ---- * To use `docker`, enter the following command: + [source,terminal,subs="+quotes,+attributes"] ---- -docker push quay.io/__/__:latest +$ docker push quay.io/__/__:latest ---- . Open your `dynamic-plugins.yaml` file to view or modify the configuration for the third-party TechDocs add-on. For example: + diff --git a/modules/upgrade/proc-upgrade-rhdh-helm.adoc b/modules/upgrade/proc-upgrade-rhdh-helm.adoc index 049e7f51fd..51e7660f05 100644 --- a/modules/upgrade/proc-upgrade-rhdh-helm.adoc +++ b/modules/upgrade/proc-upgrade-rhdh-helm.adoc @@ -34,15 +34,15 @@ It might take a few minutes to delete the resources in the older versions and to + [source,terminal,subs="+attributes,+quotes"] ---- -oc login -u -p https://api.:6443 -oc project {my-product-namespace} +$ oc login -u -p https://api.:6443 +$ oc project {my-product-namespace} ---- . For a new version of the {product-short} Helm chart, run the following upgrade command: + [source,terminal,subs="attributes+"] ---- -helm upgrade -i rhdh -f new-values.yml \ +$ helm upgrade -i rhdh -f new-values.yml \ openshift-helm-charts/redhat-developer-hub --version {product-chart-version} ---- +