This guide provides step-by-step instructions for installing the Red Hat Golden Template path using the RHEcosystemAppEng/RHDH-templates repository.
Before getting started, ensure you have the following:
- OpenShift CLI (oc): Download and install Openshift command-line interface
- Platform Access: Access to either TAP or a running RHDH instance. Helm Chart installation available here
- Hugging Face API Token: A valid authentication token from Hugging Face
Set up your Hugging Face authentication:
-
Configure your token as an environment variable:
export HF_TOKEN=<your huggingface token>
Replace with your actual Hugging Face API token.
-
Create the secret in your OpenShift namespace:
oc create secret generic huggingface-secret \ -n <your-namespace> \ --from-literal=HF_TOKEN=$HF_TOKEN
Replace with the namespace where your RAG application is deployed.
Note: Use this approach if you have Vault and External Secrets Operator configured in your cluster for centralized secret management.
-
Access Vault UI:
# Get the Vault route oc get route -n vault # Get the Vault token oc get secret -n vault vault-token -o jsonpath="{.data.token}" | base64 --decode
Open the Vault route in your browser and log in using the token method with the retrieved token.
-
Create the secret in Vault:
Note: The ExternalSecret Operator will map
hf_token
→ Kubernetes keyHF_TOKEN
-
Navigate to Create:
-
Register templates:
- Click "Register Existing Component"
-
Import the template repository:
Once registered, you'll see these AI-powered templates in the Catalog->Template page:
-
🤖 RAG Chatbot Kickstart (
chatbot-rag-kickstart-template
)
Deploy a complete RAG (Retrieval Augmented Generation) architecture using LLaMA Stack, OpenShift AI, and PGVector. Includes document ingestion pipeline and vector database for intelligent question-answering. -
🎯 AI Virtual Agent (
ai-virtual-agent-kickstart-template
)
Create an intelligent virtual assistant powered by OpenShift AI and PGVector. Perfect for building conversational AI applications with advanced reasoning capabilities. -
📊 AI Metrics Summarizer (
ai-metric-summarizer-kickstart-template
)
Build a specialized chatbot that analyzes AI model performance metrics from Prometheus and generates human-readable summaries using LLaMA models. Ideal for AI observability and monitoring.
Once you've registered the templates, follow these steps to deploy an AI application:
- From the Developer Hub sidebar, click "Create"
- You'll see the available AI templates listed
Select one of the registered templates:
- Chatbot-Rag Kickstart - for RAG document-based Q&A systems
- AI Virtual Agent - for conversational AI assistants
- AI Metrics Summarizer - for AI observability and monitoring
Fill in the guided form with your specifications:
Application Information:
- Name: Unique identifier for your component (e.g.,
my-ai-chatbot
) - Description: Brief description of your application
Repository Details:
- Host Type: Choose GitHub or GitLab
- Repository Owner: Your organization name
- Repository Name: Name for the source repository
- Namespace: Kubernetes namespace for deployment
AI Model Configuration:
- Language Model: Select from available LLaMA variants
- Safety Model: Optional LLaMA Guard for content filtering
- GPU Tolerance: Configure hardware requirements
- Review all configured parameters
- Click "Review" to validate your inputs
- Click "Create" to initiate the template deployment
The template will automatically:
- Build the software component with your specifications
- Publish source and GitOps repositories to your chosen platform
- Register the component in the Developer Hub catalog
- Deploy via ArgoCD using GitOps workflows
Once complete, use the provided links to:
- View source repository
- Monitor GitOps deployment
- Access the component in the catalog
- Review ArgoCD applications