Model Card for ViT5 Translation Model

A sequence-to-sequence translation model based on VietAI ViT5-base, fine-tuned for Vietnamese to English machine translation.
This model is intended for general-purpose translation tasks, both academic and production-oriented.


Model Details

Model Description

This model is an encoder–decoder Transformer designed for text-to-text generation tasks such as translation.
It is fine-tuned from VietAI/vit5-base, and trained in two stages:

  1. Supervised Fine-Tuning (SFT) using bilingual English–Vietnamese data
  2. Preference-based Reinforcement Learning using DPO (Direct Preference Optimization)
    to improve translation quality, fluency, and human preference alignment.
  • Developed by: tnguyen20604
  • Funded by [optional]: N/A
  • Shared by: tnguyen20604
  • Model type: Encoder–Decoder (Text-to-Text Transformer)
  • Language(s): Vietnamese, English
  • License: apache-2.0
  • Fine-tuned from: VietAI/vit5-base

Model Sources


Reinforcement Learning with DPO

After supervised fine-tuning, the model is further optimized using Direct Preference Optimization (DPO).

DPO Objective

DPO uses a dataset of chosen vs. rejected translations, allowing the model to:

  • prefer outputs closer to human preference
  • avoid unnatural translations
  • reduce hallucination and inconsistent phrasing

DPO Training Data

The preference dataset was constructed by:

  1. Generating multiple candidate translations
  2. Using a stronger LLM to score and select:
    • chosen = preferred, more fluent, more accurate translation
    • rejected = less accurate or unnatural translation
  3. Formatting into DPO tuples: { "prompt": "", "chosen": "", "rejected": "" }

Impact of DPO

  • improves fluency and human-likeness
  • reduces literal / robotic translation patterns
  • encourages coherent phrasing
  • increases BLEU and chrF scores

Uses

Direct Use

  • Machine translation Viet → Eng
  • Text rewriting (via Seq2Seq generation)
  • Academic NLP experiments
  • MT benchmarking

Downstream Use

  • Fine-tuning cho các domain đặc thù (y tế, pháp lý, kỹ thuật)
  • Tích hợp vào chatbot hoặc ứng dụng đa ngôn ngữ

Out-of-Scope Use

  • Not suitable for evaluating legal or medical content.
  • Does not guarantee full accuracy for texts containing highly specialized terminology.
  • Not appropriate for processing sensitive data or personally identifiable information (PII).

Bias, Risks, and Limitations

  • The training data originates from open-source corpora, which may introduce stylistic or domain bias.
  • The model may produce incorrect translations in cases such as:
    • ambiguous sentences
    • culturally specific expressions
    • very long or structurally complex sentences
  • Some translations may lose nuance, tone, or contextual meaning.

Recommendations

Users should manually review the translations when used in professional, safety-critical, or high-importance scenarios.


How to Get Started

from transformers import AutoTokenizer, AutoModelForSeq2SeqLM

model_name = "tnguyen20604/vit5-translation-vi2en-v2.2"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSeq2SeqLM.from_pretrained(model_name)

text = "Tôi yêu học máy."
inputs = tokenizer(text, return_tensors="pt")
outputs = model.generate(**inputs)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
1
Safetensors
Model size
0.2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for tnguyen20604/vit5-translation-vi2en-v2.2

Base model

VietAI/vit5-base
Finetuned
(88)
this model

Dataset used to train tnguyen20604/vit5-translation-vi2en-v2.2