Skip to content

Commit aa74075

Browse files
committed
fix(doc): remove comparisons
1 parent 831e185 commit aa74075

File tree

1 file changed

+0
-6
lines changed

1 file changed

+0
-6
lines changed

README.md

Lines changed: 0 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -50,12 +50,6 @@ BigCodeBench focuses on the evaluation of LLM4Code with *diverse function calls*
5050
***Precise evaluation & ranking**: See [our leaderboard](https://huggingface.co/spaces/bigcode/bigcodebench-leaderboard) for latest LLM rankings before & after rigorous evaluation.
5151
***Pre-generated samples**: BigCodeBench accelerates code intelligence research by open-sourcing [LLM-generated samples](#-LLM-generated-code) for various models -- no need to re-run the expensive benchmarks!
5252

53-
### Main Differences from EvalPlus
54-
55-
We inherit the design of the EvalPlus framework, which is a flexible and extensible evaluation framework for code generation tasks. However, BigCodeBench has the following differences:
56-
* Execution Environment: The execution environment in BigCodeBench is less bounded than EvalPlus to support tasks with diverse library dependencies.
57-
* Test Evaluation: BigCodeBench relies on `unittest` for evaluating the generated code, which is more suitable for the test harness in BigCodeBench.
58-
5953
## 🔥 Quick Start
6054

6155
> [!Tip]

0 commit comments

Comments
 (0)