Skip to content

Commit 7e92327

Browse files
Merge pull request #2147 from oracle-devrel/compute-review2025
review 2025
2 parents 93796e3 + 5a45fd1 commit 7e92327

File tree

9 files changed

+9
-9
lines changed

9 files changed

+9
-9
lines changed

cloud-infrastructure/compute-including-hpc/ai-infra-gpu/ai-infrastructure/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
This repository contains detailed information related to Oracle Cloud Infrastructure GPU compute instances.
44

5-
Reviewed: 26.02.2024
5+
Reviewed: 16.10.202
66

77
# Table of Contents
88

cloud-infrastructure/compute-including-hpc/ai-infra-gpu/ai-infrastructure/deepspeed-training-tuning/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ This repository provides a step-by-step deployment of DeepSpeed training for Lar
44

55
This setup includes a tuned DeepSpeed configuration (`tuned_ds_config.json`) that provides up to **13% performance improvement** over standard configurations.
66

7-
Reviewed: 06.06.2025
7+
Reviewed: 16.10.2025
88
# When to use this asset?
99

1010
Use this asset when you need to:

cloud-infrastructure/compute-including-hpc/ai-infra-gpu/ai-infrastructure/llm-benchmark-docker/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ and
77
using
88
[Docker Compose](https://docs.docker.com/compose/).
99

10-
Reviewed: 20.05.2025
10+
Reviewed: 16.10.2025
1111

1212
# When should this asset be used?
1313

cloud-infrastructure/compute-including-hpc/ai-infra-gpu/ai-infrastructure/nemo-megatron-training-oke/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ on the Oracle Container Engine for Kubernetes (OKE) using
88
Reference results from NVIDIA to train Llama 3 can be found on the
99
[NGC Catalog](https://catalog.ngc.nvidia.com/orgs/nvidia/teams/dgxc-benchmarking/resources/llama3-dgxc-benchmarking).
1010

11-
Reviewed: 01.07.2025
11+
Reviewed: 16.10.2025
1212

1313
# When to use this asset?
1414

cloud-infrastructure/compute-including-hpc/ai-infra-gpu/ai-infrastructure/nim-gpu-oke/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ This repository intends to demonstrate how to deploy [NVIDIA NIM](https://develo
44

55
The model used is `Llama2-7B-chat`, running on an NVIDIA A10 Tensor Core GPU hosted on OCI. For scalability, we are hosting the model repository on a Bucket in Oracle Cloud Object Storage.
66

7-
Reviewed 23.05.2024
7+
Reviewed 16.10.2025
88

99
# When to use this asset?
1010

cloud-infrastructure/compute-including-hpc/ai-infra-gpu/ai-infrastructure/rag-langchain-vllm-mistral/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
This repository is a variant of the Retrieval Augmented Generation (RAG) tutorial available [here](https://github.com/oracle-devrel/technology-engineering/tree/main/ai-and-app-modernisation/ai-services/generative-ai-service/rag-genai). Instead of the OCI GenAI Service, it uses a local deployment of Mistral 7B Instruct v0.3 using a vLLM inference server powered by an NVIDIA A10 GPU.
44

5-
Reviewed: 23.05.2024
5+
Reviewed: 16.10.2025
66

77
# When to use this asset?
88

cloud-infrastructure/compute-including-hpc/ai-infra-gpu/ai-infrastructure/triton-gpu-oke/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
This repository intends to demonstrate how to deploy NVIDIA Triton Inference Server on Oracle Kubernetes Engine (OKE) with TensorRT-LLM Backend in order to server Large Language Models (LLMs) in a Kubernetes architecture.
44

5-
Reviewed 23.05.2024
5+
Reviewed 16.10.2025
66

77
# When to use this asset?
88

cloud-infrastructure/compute-including-hpc/ai-infra-gpu/ai-infrastructure/triton-mixtral/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
This repository provides a step-by-step tutorial for deploying and using Mixtral 8x7B Large Language Model using the NVIDIA Triton Inference Server and the TensorRT-LLM backend.
44

5-
Reviewed: 23.05.2024
5+
Reviewed: 16.10.2025
66

77
# When to use this asset?
88

cloud-infrastructure/compute-including-hpc/ai-infra-gpu/ai-infrastructure/vllm-mistral/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
This repository provides a step-by-step tutorial for deploying and using [Mistral 7B Instruct](https://mistral.ai/technology/#models) Large Language Model using the [vLLM](https://github.com/vllm-project/vllm?tab=readme-ov-file) library.
44

5-
Reviewed: 23.05.2024
5+
Reviewed: 16.10.2025
66

77
# When to use this asset?
88

0 commit comments

Comments
 (0)