Skip to content

thinklis/TIMTQE

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

9 Commits
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

TIMTQE Logo

TIMTQE: Benchmarking Machine Translation Quality Estimation for Text Images

This repository provides the official code and resources for TIMTQE,
a benchmark dataset and evaluation framework for translation quality estimation (QE) on text images,
covering both synthetic (MLQE-PE) and historical (HistMTQE) settings.


πŸ“‚ Dataset

The dataset is publicly available on HuggingFace Datasets:

πŸ‘‰ https://huggingface.co/datasets/thinklis/TIMTQE

It includes:

  • MLQE-PE – large-scale synthetic subset with rendered text images.
  • HistMTQE – human-annotated historical document subset.

For detailed structure and examples, please check the HuggingFace dataset page.


βš™οΈ Evaluation

We provide an evaluation toolkit to assess the performance of quality estimation models on TIMTQE.
The main script is evaluate.py, which compares model predictions against human-annotated quality scores.

πŸ“Œ Features

  • Input Format: Predictions should be stored in a JSON, CSV, or TSV file, containing at least:

    • id (unique identifier of the sample)
    • prediction (the model’s QE score for the translation, typically on a 0–100 scale)
    • label (the human-annotated reference score)
  • Normalization: To ensure fair comparison across systems, the script applies z-score normalization to model predictions.

  • Metrics: The following evaluation metrics are computed:

    • Pearson correlation – measures the linear relationship between predictions and human scores.
    • Spearman correlation – assesses rank-based consistency between predictions and labels.
    • RMSE – penalizes larger deviations between predictions and reference scores.
    • MAE – captures the average absolute difference between predictions and labels.

πŸš€ Usage

python evaluate.py \
  --pred_file results/predictions.json \
  --ref_file data/histmtqe/test.json \
  --output_dir outputs/

πŸ“š Citation

If you use TIMTQE in your research, please cite it as follows:

@ARTICLE{11267222,
  author={Li, Shuo and Bi, Xiaojun and Sun, Yiwen},
  journal={IEEE Signal Processing Letters}, 
  title={TIMTQE: Benchmarking Machine Translation Quality Estimation for Text Images}, 
  year={2025},
  volume={},
  number={},
  pages={1-5},
  doi={10.1109/LSP.2025.3636988}
}

About

[IEEE SPL] TIMTQE: Text Image Machine Translation Quality Estimation Benchmark

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages