ComeTH (คำไทย): English-Thai Translation Quality Metrics
ComeTH is a fine-tuned version of the COMET (Crosslingual Optimized Metric for Evaluation of Translation) model specifically optimized for English-Thai translation quality assessment. This model evaluates machine translation outputs by providing quality scores that correlate highly with human judgments.
Model Overview
- Model Type: Translation Quality Estimation
- Languages: English-Thai
- Base Model: COMET (Unbabel/wmt22-cometkiwi-da)
- Encoder: XLM-RoBERTa-based (microsoft/infoxlm-large)
- Architecture: Unified Metric with sentence-level scoring
- Framework: COMET (Unbabel)
- Task: Machine Translation Evaluation
- Parameters: 565M (558M encoder + 6.3M estimator)
Versions
We offer two variants of ComeTH with different training approaches:
- ComeTH: Fine-tuned on human MQM annotations (Spearman's ρ = 0.4639)
- ComeTH-Augmented: Fine-tuned on human + Claude-assisted annotations (Spearman's ρ = 0.4795)
Both models outperform the base COMET model (Spearman's ρ = 0.4570) on English-Thai translation evaluation. The Claude-augmented version leverages LLM-generated annotations to enhance correlation with human judgments by 4.9% over the baseline.
Technical Specifications
- Training Framework: PyTorch Lightning
- Loss Function: MSE
- Input Segments: [mt, src]
- Final Layer Architecture: [3072, 1024]
- Layer Transformation: Sparsemax
- Activation Function: Tanh
- Dropout: 0.1
- Learning Rate: 1.5e-05 (Encoder: 1e-06)
- Layerwise Decay: 0.95
- Word Layer: 24
Training Data
The models were trained on:
- Size: 23,530 English-Thai translation pairs
- Source Domains: Diverse, including technical, conversational, and e-commerce
- Annotation Framework: Multidimensional Quality Metrics (MQM)
- Error Categories:
- Minor: Issues that don't significantly impact meaning or usability
- Major: Errors that significantly impact meaning but don't render content unusable
- Critical: Errors that make content unusable or could have serious consequences
- Claude Augmentation: Claude 3.5 Sonnet was used to generate supplementary quality judgments, enhancing the model's alignment with human evaluations
Training Process
ComeTH was trained using a multi-step process:
- Starting from the wmt22-cometkiwi-da checkpoint
- Fine-tuning on human MQM annotations for 5 epochs
- Using gradient accumulation (8 steps) to simulate larger batch sizes
- Utilizing unified metric architecture that combines source and MT embeddings
- For the augmented variant: additional training with Claude-assisted annotations, weighted to balance human and machine judgments
Performance
Correlation with Human Judgments (Spearman's ρ)
Model | Spearman's ρ | RMSE |
---|---|---|
COMET (baseline) | 0.4570 | 0.3185 |
ComeTH (human annotations) | 0.4639 | 0.3093 |
ComeTH-Augmented (human + Claude) | 0.4795 | 0.3078 |
The Claude-augmented version demonstrates the highest correlation with human judgments, offering a significant improvement over both the baseline and human-only models.
Comparison with Other LLM Evaluators
Model | Spearman's ρ |
---|---|
ComeTH-Augmented | 0.4795 |
Claude 3.5 Sonnet | 0.4383 |
GPT-4o Mini | 0.4352 |
Gemini 2.0 Flash | 0.3918 |
ComeTH-Augmented outperforms direct evaluations from state-of-the-art LLMs, while being more computationally efficient for large-scale translation quality assessments.
Advanced Usage Examples
Basic Evaluation
from comet import download_model, load_from_checkpoint
model_path = download_model("wasanx/ComeTH")
model = load_from_checkpoint(model_path)
translations = [
{
"src": "This is an English source text.",
"mt": "นี่คือข้อความภาษาอังกฤษ",
}
]
results = model.predict(translations, batch_size=8, gpus=1)
scores = results['scores']
Batch Processing With Progress Tracking
import pandas as pd
from tqdm import tqdm
df = pd.read_csv("translations.csv")
input_data = df[['src', 'mt']].to_dict('records')
batch_size = 32
all_scores = []
for i in tqdm(range(0, len(input_data), batch_size)):
batch = input_data[i:i+batch_size]
results = model.predict(batch, batch_size=len(batch), gpus=1)
all_scores.extend(results['scores'])
df['quality_score'] = all_scores
System-Level Evaluation
import numpy as np
systems = df.groupby('system_name')['quality_score'].agg(['mean', 'std', 'count']).reset_index()
systems = systems.sort_values('mean', ascending=False)
print(systems)
Citation
@misc{
title = {COMETH: English-Thai Translation Quality Metrics},
author = {COMETH Team},
year = {2025},
howpublished = {Hugging Face Model Repository},
url = {https://huggingface.co/wasanx/ComeTH}
}
Contact
For questions or feedback: [email protected]
License
The COMETH Reserved License
Cometh English-to-Thai Translation Data and Model License
Copyright (C) Cometh Team. All rights reserved.
This license governs the use of the Cometh English-to-Thai translation data and model ("Cometh Model Data"), including but not limited to MQM scores, human translations, and human rankings from various translation sources.
Permitted Use
The Cometh Model Data is licensed exclusively for internal use by the designated Cometh team.
Prohibited Use
The following uses are strictly prohibited:
1. Any usage outside the designated purposes unanimously approved by the Cometh team.
2. Redistribution, sharing, or distribution of the Cometh Model Data in any form.
3. Citation or public reference to the Cometh Model Data in any academic, commercial, or non-commercial context.
4. Any use beyond the internal operations of the Cometh team.
Legal Enforcement
Unauthorized use, distribution, or citation of the Cometh Model Data constitutes a violation of this license and may result in legal action, including but not limited to prosecution under applicable laws.
Reservation of Rights
All rights to the Cometh Model Data are reserved by the Cometh team. This license does not transfer any ownership rights.
By accessing or using the Cometh Model Data, you agree to be bound by the terms of this license.
Model tree for wasanx/ComeTH
Datasets used to train wasanx/ComeTH
Space using wasanx/ComeTH 1
Collection including wasanx/ComeTH
Evaluation results
- Spearman correlation on COMETH Claude Augmentation Datasetsself-reported0.479
- COMET baseline on COMETH Baseline Comparisonself-reported0.457
- ComETH (human-only) on COMETH Baseline Comparisonself-reported0.464