Skip to content

Commit 38e7664

Browse files
author
Alberto Cabrera
committed
Reverting llama2 -> llama3 changes
1 parent 501891c commit 38e7664

File tree

1 file changed

+20
-10
lines changed

1 file changed

+20
-10
lines changed

docs/backend/SYCL.md

Lines changed: 20 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -353,7 +353,7 @@ cmake --build build --config Release -j -v
353353

354354
#### Retrieve and prepare model
355355

356-
You can refer to the general [*Prepare and Quantize*](README.md#prepare-and-quantize) guide for model prepration, or download an already quantized model like [Meta-Llama-3-8B-Instruct-Q4_0.gguf](https://huggingface.co/aptha/Meta-Llama-3-8B-Instruct-Q4_0-GGUF/resolve/main/Meta-Llama-3-8B-Instruct-Q4_0.gguf).
356+
You can refer to the general [*Prepare and Quantize*](README.md#prepare-and-quantize) guide for model preparation, or download an already quantized model like [llama-2-7b.Q4_0.gguf](https://huggingface.co/TheBloke/Llama-2-7B-GGUF/blob/main/llama-2-7b.Q4_0.gguf) or [Meta-Llama-3-8B-Instruct-Q4_0.gguf](https://huggingface.co/aptha/Meta-Llama-3-8B-Instruct-Q4_0-GGUF/resolve/main/Meta-Llama-3-8B-Instruct-Q4_0.gguf).
357357

358358
##### Check device
359359

@@ -399,12 +399,16 @@ Choose one of following methods to run.
399399
- Use device 0:
400400

401401
```sh
402-
./examples/sycl/run-llama.sh 0
402+
./examples/sycl/run-llama2.sh 0
403+
# OR
404+
./examples/sycl/run-llama3.sh 0
403405
```
404406
- Use multiple devices:
405407

406408
```sh
407-
./examples/sycl/run-llama.sh
409+
./examples/sycl/run-llama2.sh
410+
# OR
411+
./examples/sycl/run-llama3.sh
408412
```
409413

410414
2. Command line
@@ -427,13 +431,13 @@ Examples:
427431
- Use device 0:
428432

429433
```sh
430-
ZES_ENABLE_SYSMAN=1 ./build/bin/llama-cli -no-cnv -m models/Meta-Llama-3-8B-Instruct-Q4_0.gguf -p "Building a website can be done in 10 simple steps:" -n 400 -e -ngl 33 -sm none -mg 0
434+
ZES_ENABLE_SYSMAN=1 ./build/bin/llama-cli -no-cnv -m models/llama-2-7b.Q4_0.gguf -p "Building a website can be done in 10 simple steps:" -n 400 -e -ngl 99 -sm none -mg 0
431435
```
432436

433437
- Use multiple devices:
434438

435439
```sh
436-
ZES_ENABLE_SYSMAN=1 ./build/bin/llama-cli -no-cnv -m models/Meta-Llama-3-8B-Instruct-Q4_0.gguf -p "Building a website can be done in 10 simple steps:" -n 400 -e -ngl 33 -sm layer
440+
ZES_ENABLE_SYSMAN=1 ./build/bin/llama-cli -no-cnv -m models/llama-2-7b.Q4_0.gguf -p "Building a website can be done in 10 simple steps:" -n 400 -e -ngl 99 -sm layer
437441
```
438442

439443
*Notes:*
@@ -454,7 +458,7 @@ use 1 SYCL GPUs: [0] with Max compute units:512
454458

455459
1. Install GPU driver
456460

457-
Intel GPU drivers instructions guide and download page can be found here: [Get intel GPU Drivers](https://www.intel.com/content/www/us/en/products/docs/discrete-gpus/arc/software/drivers.html).
461+
Intel GPU drivers instructions guide and download page can be found here: [Get Intel GPU Drivers](https://www.intel.com/content/www/us/en/products/docs/discrete-gpus/arc/software/drivers.html).
458462

459463
2. Install Visual Studio
460464

@@ -631,7 +635,7 @@ Once it is completed, final results will be in **build/Release/bin**
631635

632636
#### Retrieve and prepare model
633637

634-
You can refer to the general [*Prepare and Quantize*](README.md#prepare-and-quantize) guide for model prepration, or simply download [Meta-Llama-3-8B-Instruct-Q4_0.gguf](https://huggingface.co/aptha/Meta-Llama-3-8B-Instruct-Q4_0-GGUF/blob/main/Meta-Llama-3-8B-Instruct-Q4_0.gguf) model as example.
638+
You can refer to the general [*Prepare and Quantize*](README.md#prepare-and-quantize) guide for model preparation, or download an already quantized model like [llama-2-7b.Q4_0.gguf](https://huggingface.co/TheBloke/Llama-2-7B-GGUF/blob/main/llama-2-7b.Q4_0.gguf) or [Meta-Llama-3-8B-Instruct-Q4_0.gguf](https://huggingface.co/aptha/Meta-Llama-3-8B-Instruct-Q4_0-GGUF/resolve/main/Meta-Llama-3-8B-Instruct-Q4_0.gguf).
635639

636640
##### Check device
637641

@@ -676,7 +680,13 @@ Choose one of following methods to run.
676680
1. Script
677681

678682
```
679-
examples\sycl\win-run-llama.bat
683+
examples\sycl\win-run-llama-2.bat
684+
```
685+
686+
or
687+
688+
```
689+
examples\sycl\win-run-llama-3.bat
680690
```
681691

682692
2. Command line
@@ -700,13 +710,13 @@ Examples:
700710
- Use device 0:
701711

702712
```
703-
build\bin\llama-cli.exe -no-cnv -m models\Meta-Llama-3-8B-Instruct-Q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e -ngl 99 -sm none -mg 0
713+
build\bin\llama-cli.exe -no-cnv -m models\llama-2-7b.Q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e -ngl 99 -sm none -mg 0
704714
```
705715

706716
- Use multiple devices:
707717

708718
```
709-
build\bin\llama-cli.exe -no-cnv -m models\Meta-Llama-3-8B-Instruct-Q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e -ngl 99 -sm layer
719+
build\bin\llama-cli.exe -no-cnv -m models\llama-2-7b.Q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e -ngl 99 -sm layer
710720
```
711721

712722

0 commit comments

Comments
 (0)