Skip to content

Commit d4696e1

Browse files
authored
Merge pull request #219 from e06084/dev
feat: update metric doc
2 parents 8dded23 + d818ca3 commit d4696e1

14 files changed

+30
-28
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -197,7 +197,7 @@ Dingo provides comprehensive data quality assessment through both rule-based and
197197
📊 **[View Complete Metrics Documentation →](docs/metrics.md)**
198198

199199
Our evaluation system includes:
200-
- **Text Quality Assessment Metrics**: Pre-training data quality evaluation using DataMan methodology and enhanced multi-dimensional assessment
200+
- **Pretrain Text Quality Assessment Metrics**: Pre-training data quality evaluation using DataMan methodology and enhanced multi-dimensional assessment
201201
- **SFT Data Assessment Metrics**: Honest, Helpful, Harmless evaluation for supervised fine-tuning data
202202
- **Classification Metrics**: Topic categorization and content classification
203203
- **Multimodality Assessment Metrics**: Image classification and relevance evaluation

dingo/model/llm/llm_factcheck_public.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@ class LLMFactCheckPublic(BaseOpenAI):
3232
"""公开事实性评估器 - 基于 GPT-5 System Card 的两阶段评估"""
3333

3434
_metric_info = {
35-
"category": "Factuality Assessment",
35+
"category": "SFT Data Assessment Metrics",
3636
"quality_dimension": "FACTUAL_CORRECTNESS",
3737
"metric_name": "LLMFactCheckPublic",
3838
"description": "Two-stage factuality evaluation pipeline from GPT-5",

dingo/model/prompt/prompt_code_compare.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
@Model.prompt_register('CodeCompare', [], ['LLMCodeCompare'])
66
class PromptCodeCompare(BasePrompt):
77
_metric_info = {
8-
'category': 'SFT Data Assessment Metrics',
8+
'category': 'Pretrain Text Quality Assessment Metrics',
99
'metric_name': 'PromptCodeCompare',
1010
'description': 'Compares the effectiveness of two tools in extracting code blocks from HTML to Markdown format by evaluating recognition rate and accuracy to determine which tool performs better',
1111
'paper_title': '',

dingo/model/prompt/prompt_dataman_assessment.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -87,7 +87,7 @@ class PromptDataManAssessment(BasePrompt):
8787

8888
# Metadata for documentation generation
8989
_metric_info = {
90-
"category": "Text Quality Assessment Metrics",
90+
"category": "Pretrain Text Quality Assessment Metrics",
9191
"metric_name": "PromptDataManAssessment",
9292
"description": "Evaluates pre-training data quality using the DataMan methodology (14 standards, 15 domains). Assigns a score (0/1), domain type, quality status, and reason.",
9393
"paper_title": "DataMan: Data Manager for Pre-training Large Language Models",

dingo/model/prompt/prompt_document_parsing.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66
class PromptDocumentParsingQuality(BasePrompt):
77
# Metadata for documentation generation
88
_metric_info = {
9-
"category": "Document Parsing",
9+
"category": "OCR Eval Metric",
1010
"metric_name": "PromptDocumentParsingQuality",
1111
"description": "Evaluate the quality of general document parsing",
1212
"evaluation_results": "",

dingo/model/prompt/prompt_html_extract_compare.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
@Model.prompt_register("Html_Extract_Compare", [], ['LLMHtmlExtractCompare'])
66
class PromptHtmlExtractCompare(BasePrompt):
77
_metric_info = {
8-
'category': 'SFT Data Assessment Metrics',
8+
'category': 'Pretrain Text Quality Assessment Metrics',
99
'metric_name': 'PromptHtmlExtractCompare',
1010
'description': 'Compares the effectiveness of two HTML extraction tools by evaluating element recognition rate and accuracy across different content types',
1111
'paper_title': '',

dingo/model/prompt/prompt_html_extract_compare_v2.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
@Model.prompt_register("Html_Extract_Compare_V2", ['html_extract_compare'], ['LLMHtmlExtractCompareV2'])
66
class PromptHtmlExtractCompareV2(BasePrompt):
77
_metric_info = {
8-
'category': 'SFT Data Assessment Metrics',
8+
'category': 'Pretrain Text Quality Assessment Metrics',
99
'metric_name': 'PromptHtmlExtractCompareV2',
1010
'description': 'Compares HTML extraction results using diff-match-patch algorithm to identify unique and common content, then evaluates core informational content differences',
1111
'paper_title': '',

dingo/model/prompt/prompt_math_compare.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
@Model.prompt_register('MathCompare', [], ['LLMMathCompare'])
66
class PromptMathCompare(BasePrompt):
77
_metric_info = {
8-
'category': 'SFT Data Assessment Metrics',
8+
'category': 'Pretrain Text Quality Assessment Metrics',
99
'metric_name': 'PromptMathCompare',
1010
'description': 'Compares the effectiveness of two tools in extracting mathematical formulas from HTML to Markdown format by evaluating recognition rate and accuracy to determine which tool performs better',
1111
'paper_title': '',

dingo/model/prompt/prompt_politics.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ class PromptPolitics(BasePrompt):
77

88
# Metadata for documentation generation
99
_metric_info = {
10-
"category": "Text Quality Assessment Metrics",
10+
"category": "Pretrain Text Quality Assessment Metrics",
1111
"metric_name": "PromptPolitics",
1212
"description": "Evaluates whether the text contains politics-related content",
1313
"paper_title": "",

dingo/model/prompt/prompt_table_compare.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
@Model.prompt_register('TableCompare', [], ['LLMTableCompare'])
66
class PromptTableCompare(BasePrompt):
77
_metric_info = {
8-
'category': 'SFT Data Assessment Metrics',
8+
'category': 'Pretrain Text Quality Assessment Metrics',
99
'metric_name': 'PromptTableCompare',
1010
'description': 'Compares the effectiveness of two tools in extracting tables from HTML to Markdown format by evaluating recognition rate and accuracy to determine which tool performs better',
1111
'paper_title': '',

0 commit comments

Comments
 (0)