Skip to content

Commit ef1f4c8

Browse files
Update tools/benchmarks/llm_eval_harness/meta_eval_reproduce/README.md
Co-authored-by: Hamid Shojanazeri <[email protected]>
1 parent c32517b commit ef1f4c8

File tree

1 file changed

+1
-1
lines changed
  • tools/benchmarks/llm_eval_harness/meta_eval_reproduce

1 file changed

+1
-1
lines changed

tools/benchmarks/llm_eval_harness/meta_eval_reproduce/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ As Meta Llama models gain popularity, evaluating these models has become increas
66
## Disclaimer
77

88

9-
1. **This tutorial is not the official implementation** of Meta Llama evaluation. It is based on public third-party libraries, and the implementation may differ slightly from our internal evaluation, leading to minor differences in the reproduced numbers.
9+
1. **This recipe is not the official implementation** of Meta Llama evaluation. It is based on public third-party libraries, as this implementation is not mirroring Meta Llama evaluation, this may lead to minor differences in the reproduced numbers.
1010
2. **Model Compatibility**: This tutorial is specifically for Llama 3 based models, as our prompts include Meta Llama 3 special tokens, e.g. `<|start_header_id|>user<|end_header_id|>`. It will not work with models that are not based on Llama 3.
1111

1212
### Differences between our evaluation and Hugging Face leaderboard evaluation

0 commit comments

Comments
 (0)