Fine-tuned mBART Model for English to Urdu Translation
This repository contains a fine-tuned mBART model for English to Urdu translation. The model has been trained on a custom dataset and evaluated on test data.
Model Information
- Model Name:
abdulwaheed1/english-to-urdu-translation-mbart
- Base Model:
facebook/mbart-large-50
- Tokenizer:
facebook/mbart-large-50
- Source Language: English (
en
) - Target Language: Urdu (
ur
)
Usage
python
from transformers import MBart50TokenizerFast, MBartForConditionalGeneration
# Load the fine-tuned model
model_name = "abdulwaheed1/english-to-urdu-translation-mbart"
tokenizer = MBart50TokenizerFast.from_pretrained(model_name, src_lang="en_XX", tgt_lang="ur_PK")
model = MBartForConditionalGeneration.from_pretrained(model_name)
Evaluation
The model has been evaluated on a test dataset, and the following metrics were obtained:
- BLEU Score: 35.87
- Generation Length: 42.56
- Meteor Score: 0.60
Training Details
The model was trained using the transformers
library with the following configuration:
- Training Loss: 1.5697
- Validation Loss: 1.1256
Dataset
The model was fine-tuned on a custom English-Urdu translation dataset. If you wish to use the same dataset, you can find the preprocessing script and dataset files in the data
directory.
Acknowledgments
The fine-tuning process and code were inspired by the Hugging Face Transformers library.
- Downloads last month
- 92
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.