barbaroo/nllb_200_600M_en_fo

Model Description

  • Model Architecture: This model is based on the NLLB 600M architecture and weights.
  • Languages: This checkpoint is fine-tuned to translate from English (en) to Faroese (fo).
  • Size: ~600M parameters.
  • Finetuning Datasets:
  • License: Inherits the original licenses of the NLLB 600M model.

Intended Use

  • Primary Use Case: Translate text from English to Faroese.
  • Audience: Researchers, developers, or anyone interested in Faroese language processing.
  • Usage Scenarios:
    • Building Faroese-English translation tools
    • Language research and corpus analysis
    • Synthetic data creation

Important: While the model can produce fluent translations, it is not guaranteed to be perfectly accurate on all inputs. Users should verify critical or sensitive content through human experts.

Metrics

  • Model performance measures:
    NLLB-200 model was evaluated using BLEU, chrF and BERT-score —metrics widely adopted by the machine translation community.

Evaluation Data

  • Datasets:
    Flores-200 dataset is described in Section 4 of the NLLB paper/documentation.

  • Motivation:
    Flores-200 is currently the only machine translation benchmark available for Faroese.

    How to Use

Below is a simple usage example in Python with Hugging Face Transformers:

from transformers import pipeline

model_name = "barbaroo/nllb_200_600M_en_fo"

translator = pipeline(
    "translation",
    model=model_name,
    tokenizer=model_name,
    src_lang="eng_Latn",   # Language code for English
    tgt_lang="fao_Latn"    # Language code for Faroese
)

text = "Hello, how are you?"
translation = translator(text)
print(translation)

Citation

If you use this model or find it helpful in your research, please cite: [COMING SOON]

Contact

For questions, feedback, or collaboration inquiries, feel free to reach out:

Downloads last month
24
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for barbaroo/nllb_200_600M_en_fo

Finetuned
(98)
this model

Datasets used to train barbaroo/nllb_200_600M_en_fo