This documentation outlines the deployment procedures for the Peridot kernel on non-NVIDIA architectures, specifically AMD Radeon and Intel Arc graphics processors.
These configurations are strictly community-maintained.
The Peridot core architecture is optimized and officially validated exclusively on NVIDIA environments; therefore, alternative GPU deployments require manual configuration and are provided without official guarantees of stability.
Hardware-specific issues should be logged in the repository's Issue tracker utilizing the [AMD] or [Intel Arc] tags.
Support for AMD graphics processing units is currently restricted to Linux environments utilizing the ROCm framework.
Windows deployments via ROCm are highly experimental and are not supported within this documentation.
| Architecture | VRAM | Validation Status | Expected Inference (t/s) |
|---|---|---|---|
| RX 6600 | 8GB | Confirmed | 35-42 |
| RX 6700 XT | 12GB | Confirmed | 42-50 |
| RX 6800 XT | 16GB | Confirmed | 48-55 |
| RX 6900 XT | 16GB | Confirmed | 50-58 |
| RX 7600 | 8GB | Limited Testing | 38-45 (Est.) |
| RX 7700 XT | 12GB | Limited Testing | 45-52 (Est.) |
| RX 7800 XT | 16GB | Confirmed | 52-60 |
| RX 7900 XT | 20GB | Confirmed | 58-65 |
| RX 7900 XTX | 24GB | Confirmed | 60-70 |
- Operating System: Ubuntu 22.04 LTS, Ubuntu 24.04 LTS, or Debian 12
- Kernel: 5.15 or later
- Drivers: AMDGPU kernel module
- Environment: Python 3.11
Ensure the operating system recognizes the AMD architecture:
lspci | grep -i amd
lsmod | grep amdgpuExecute the following commands based on your Linux distribution to install the Radeon Open Compute (ROCm) stack.
wget https://repo.radeon.com/amdgpu-install/6.0/ubuntu/noble/amdgpu-install_6.0.60000-1_all.deb
sudo dpkg -i amdgpu-install_6.0.60000-1_all.deb
sudo apt update
sudo amdgpu-install --usecase=rocm
sudo usermod -a -G render,video $USER
sudo rebootwget https://repo.radeon.com/amdgpu-install/6.0/ubuntu/jammy/amdgpu-install_6.0.60000-1_all.deb
sudo dpkg -i amdgpu-install_6.0.60000-1_all.deb
sudo apt update
sudo amdgpu-install --usecase=rocm
sudo usermod -a -G render,video $USER
sudo rebootPost-reboot, confirm the compute nodes are active:
/opt/rocm/bin/rocminfo
/opt/rocm/bin/rocm-smigit clone https://github.com/uncoalesced/Peridot.git
cd Peridot
python3.11 -m venv venv
source venv/bin/activate
pip install --upgrade pipThe llama-cpp-python binding must be compiled manually to interface with the ROCm backend.
export ROCM_PATH=/opt/rocm
export HIP_PATH=/opt/rocm
sudo apt install cmake build-essential
CMAKE_ARGS="-DLLAMA_HIPBLAS=on" pip install llama-cpp-python --no-cache-dir --force-reinstall --upgradepip install -r requirements.txt
mkdir -p models
wget -O models/llama-3-8b-q4.gguf https://huggingface.co/QuantFactory/Meta-Llama-3-8B-Instruct-GGUF/resolve/main/Meta-Llama-3-8B-Instruct.Q4_K_M.ggufModify the configuration file to address the AMD hardware:
MODEL_PATH = "models/llama-3-8b-q4.gguf"
GPU_LAYERS = 33 # Adjust downward if VRAM is <16GB
N_CTX = 4096
USE_GPU = True
GPU_TYPE = "amd"export GPU_DEVICE_ORDINAL=0
export HIP_VISIBLE_DEVICES=0
python launcher.pyIntel Arc integrations utilize the Vulkan backend, which maintains operational compatibility across both Windows 11 and Linux environments.
| Architecture | VRAM | Validation Status | Expected Inference (t/s) |
|---|---|---|---|
| Arc A310 | 4GB | Limited Testing | 15-20 |
| Arc A380 | 6GB | Confirmed | 20-28 |
| Arc A750 | 8GB | Confirmed | 25-35 |
| Arc A770 | 8GB | Confirmed | 28-38 |
| Arc A770 | 16GB | Confirmed | 30-42 |
Windows:
Install the latest Intel Arc graphics drivers and the LunarG Vulkan SDK (Runtime).
Linux (Ubuntu 22.04+):
Ensure kernel 6.2+ is active. Install necessary runtime packages:
sudo apt update
sudo apt install intel-opencl-icd vulkan-tools mesa-vulkan-driversEnsure the virtual environment is active before compilation.
$env:CMAKE_ARGS="-DLLAMA_VULKAN=on"
pip install llama-cpp-python --no-cache-dir --force-reinstall --upgradesudo apt install cmake build-essential
CMAKE_ARGS="-DLLAMA_VULKAN=on" pip install llama-cpp-python --no-cache-dir --force-reinstall --upgradeAdjust the VRAM layer offloading to accommodate Intel architecture:
MODEL_PATH = "models/llama-3-8b-q4.gguf"
GPU_LAYERS = 25 # Recommended baseline for 8GB Arc GPUs
N_CTX = 4096
USE_GPU = True
GPU_TYPE = "intel_arc"The Peridot kernel's VRAM handoff for medical research (Folding@home) natively supports both AMD (via OpenCL) and Intel (via Vulkan) backends.
To initialize the daemon on either architecture, run:
python medical_research.py setupAMD RX 6600+ Expected Throughput: ~300,000 PPD (Points Per Day).
Intel Arc A750+ Expected Throughput: ~200,000 PPD.
To expand our hardware validation matrices, community members are encouraged to submit performance telemetry.
If you achieve stable execution on undocumented hardware, please fork this repository, update the relevant markdown tables, and submit a Pull Request formatted as follows:
Hardware Telemetry: [GPU Model]
- Hardware Architecture
- OS Version / Kernel
- Driver or ROCm Build Version
- Sustained Inference Speed (t/s)
- Layer Configuration (GPU_LAYERS)
Peridot dynamically scales its Retrieval-Augmented Generation (RAG) capabilities based on available VRAM to prevent Out-Of-Memory (OOM) faults and preserve the 21ms Folding@home handoff latency.
- Target Hardware: Intel Arc A310, AMD RX 6500 XT, older community hardware.
- Embedding Model:
all-MiniLM-L3-v2(~45MB). - Execution: Forced strictly to CPU.
- Details: Leaves 100% of the limited VRAM available for the quantized LLM. Retrieval takes slightly longer, but system stability is guaranteed.
- Target Hardware: AMD RX 6600, Intel Arc A750.
- Embedding Model:
all-MiniLM-L6-v2(~90MB). - Execution: Forced strictly to CPU.
- Details: The standard Peridot configuration. Balances highly accurate retrieval with zero VRAM footprint, protecting the background medical research state machine.
- Target Hardware: AMD RX 6700 XT, RX 7800 XT, Intel Arc A770.
- Embedding Model:
nomic-embed-text-v1.5(~550MB). - Execution: VRAM Accelerated.
- Details: Utilizes the VRAM buffer to load a massive 8192-token context window. Capable of ingesting entire document chapters in a single pass with sub-millisecond retrieval.
- Target Hardware: AMD RX 7900 XTX.
- Embedding Model:
mxbai-embed-large-v1(~1.5GB). - Execution: VRAM Accelerated.
- Details: Enterprise-grade retrieval. Best suited for massive personal databases and deep semantic search operations.
For technical assistance regarding these implementations, utilize the GitHub Discussions panel or open a properly tagged Issue.