Skip to content

Commit edab9cd

Browse files
pagezyhfjulien-c
andauthored
model_name raising error because of the dot (#1459)
* model_name raising error because of the . * Update docs/sagemaker/inference.md Co-authored-by: Julien Chaumond <[email protected]> --------- Co-authored-by: Julien Chaumond <[email protected]>
1 parent 88ffdc5 commit edab9cd

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

docs/sagemaker/inference.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -358,12 +358,12 @@ You should also define `SM_NUM_GPUS`, which specifies the tensor parallelism deg
358358
Note that you can optionally reduce the memory and computational footprint of the model by setting the `HF_MODEL_QUANTIZE` environment variable to `true`, but this lower weight precision could affect the quality of the output for some models.
359359

360360
```python
361-
model_name = "llama-3.1-8b-instruct" + time.strftime("%Y-%m-%d-%H-%M-%S", time.gmtime())
361+
model_name = "llama-3-1-8b-instruct" + time.strftime("%Y-%m-%d-%H-%M-%S", time.gmtime())
362362

363363
hub = {
364-
'HF_MODEL_ID':'EleutherAI/gpt-neox-20b',
364+
'HF_MODEL_ID':'meta-llama/Llama-3.1-8B-Instruct',
365365
'SM_NUM_GPUS':'1',
366-
'HUGGING_FACE_HUB_TOKEN': '<REPLACE WITH YOUR TOKEN>'
366+
'HUGGING_FACE_HUB_TOKEN': '<REPLACE WITH YOUR TOKEN>',
367367
}
368368

369369
assert hub['HUGGING_FACE_HUB_TOKEN'] != '<REPLACE WITH YOUR TOKEN>', "You have to provide a token."

0 commit comments

Comments
 (0)