Skip to content

Commit 971236c

Browse files
authored
Merge pull request #1534 from hanhainebula/master
Support evaluation: BRIGHT
2 parents c56fa29 + 545f0d7 commit 971236c

File tree

10 files changed

+831
-4
lines changed

10 files changed

+831
-4
lines changed
Lines changed: 17 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,17 @@
1+
from FlagEmbedding.abc.evaluation import (
2+
AbsEvalModelArgs as BrightEvalModelArgs,
3+
)
4+
5+
from .data_loader import BrightShortEvalDataLoader, BrightLongEvalDataLoader
6+
from .arguments import BrightEvalArgs
7+
from .runner import BrightEvalRunner
8+
from .searcher import BrightEvalDenseRetriever
9+
10+
__all__ = [
11+
"BrightEvalArgs",
12+
"BrightEvalModelArgs",
13+
"BrightEvalRunner",
14+
"BrightEvalDenseRetriever",
15+
"BrightShortEvalDataLoader",
16+
"BrightLongEvalDataLoader",
17+
]
Lines changed: 28 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,28 @@
1+
from transformers import HfArgumentParser
2+
3+
from FlagEmbedding.evaluation.bright import (
4+
BrightEvalArgs, BrightEvalModelArgs,
5+
BrightEvalRunner
6+
)
7+
8+
9+
def main():
10+
parser = HfArgumentParser((
11+
BrightEvalArgs,
12+
BrightEvalModelArgs
13+
))
14+
15+
eval_args, model_args = parser.parse_args_into_dataclasses()
16+
eval_args: BrightEvalArgs
17+
model_args: BrightEvalModelArgs
18+
19+
runner = BrightEvalRunner(
20+
eval_args=eval_args,
21+
model_args=model_args
22+
)
23+
24+
runner.run()
25+
26+
27+
if __name__ == "__main__":
28+
main()
Lines changed: 16 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,16 @@
1+
from dataclasses import dataclass, field
2+
3+
from FlagEmbedding.abc.evaluation.arguments import AbsEvalArgs
4+
5+
6+
@dataclass
7+
class BrightEvalArgs(AbsEvalArgs):
8+
"""
9+
Argument class for Bright evaluation.
10+
"""
11+
task_type: str = field(
12+
default="short", metadata={"help": "The task type to evaluate on. Available options: ['short', 'long']. Default: short", "choices": ["short", "long"]}
13+
)
14+
use_special_instructions: bool = field(
15+
default=True, metadata={"help": "Whether to use specific instructions in `prompts.py` for evaluation. Default: True"}
16+
)

0 commit comments

Comments
 (0)