Skip to content

Commit 86ae713

Browse files
Update Readme.md
1 parent 53c2a03 commit 86ae713

File tree

1 file changed

+10
-10
lines changed
  • cloud-infrastructure/ai-infra-gpu/GPU/triton-gpu-oke

1 file changed

+10
-10
lines changed

cloud-infrastructure/ai-infra-gpu/GPU/triton-gpu-oke/README.md

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -4,11 +4,11 @@ This repository intends to demonstrate how to deploy NVIDIA Triton Inference Ser
44

55
# Pre-requisites
66

7-
You have access to an Oracle Cloud Tenancy.
8-
You have access to shapes with NVIDIA GPU such as A10 GPU's (i.e VM.GPU.A10.1).
9-
You have a [container registry](https://docs.oracle.com/en-us/iaas/Content/Registry/home.htm).
10-
You have an [Auth Token](https://docs.oracle.com/en-us/iaas/Content/Registry/Tasks/registrypushingimagesusingthedockercli.htm#Pushing_Images_Using_the_Docker_CLI) to push/pull images to/from the registry.
11-
You are familiar with Kubernetes and Helm basic terminology.
7+
* You have access to an Oracle Cloud Tenancy.
8+
* You have access to shapes with NVIDIA GPU such as A10 GPU's (i.e VM.GPU.A10.1).
9+
* You have a [container registry](https://docs.oracle.com/en-us/iaas/Content/Registry/home.htm).
10+
* You have an [Auth Token](https://docs.oracle.com/en-us/iaas/Content/Registry/Tasks/registrypushingimagesusingthedockercli.htm#Pushing_Images_Using_the_Docker_CLI) to push/pull images to/from the registry.
11+
* You are familiar with Kubernetes and Helm basic terminology.
1212

1313
# Walkthrough
1414

@@ -215,7 +215,7 @@ sudo docker push <region-key>.ocir.io/<tenancy-namespace>/triton_llm:triton_trt_
215215

216216
Here is the target architecture at the end of the deployment:
217217

218-
[Architecture Diagram](architecture-diagram.png)
218+
![Architecture Diagram](architecture-diagram.png)
219219

220220
At this stage, your container is ready and uploaded to your Oracle Registry. It is now time to bring everything together in Oracle Kubernetes Engines (OKE)
221221

@@ -367,9 +367,9 @@ $ oci os bucket delete --bucket-name triton-inference-server-repository --empty
367367

368368
Resources:
369369

370-
[TensorRT-LLM Backend on Triton official repo](https://github.com/triton-inference-server/tensorrtllm_backend)
371-
[NCG page with all version of NVIDIA Triton Inference Server](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/tritonserver/tags)
372-
[LLaMA Example](https://github.com/NVIDIA/TensorRT-LLM/tree/e06f537e08f792fd52e6fef7bbc7b54774492503/examples/llama)
373-
[Triton deployment on Cloud Provider](https://github.com/triton-inference-server/server/tree/main/deploy)
370+
* [TensorRT-LLM Backend on Triton official repo](https://github.com/triton-inference-server/tensorrtllm_backend)
371+
* [NCG page with all version of NVIDIA Triton Inference Server](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/tritonserver/tags)
372+
* [LLaMA Example](https://github.com/NVIDIA/TensorRT-LLM/tree/e06f537e08f792fd52e6fef7bbc7b54774492503/examples/llama)
373+
* [Triton deployment on Cloud Provider](https://github.com/triton-inference-server/server/tree/main/deploy)
374374

375375

0 commit comments

Comments
 (0)