bert-mrpc-finetuned / README.md
dcarpintero's picture
Update README.md
c61ec0f verified
metadata
language: en
tags:
  - bert
  - sequence-classification
  - mrpc
  - paraphrase
license: mit

Model description

Fine-tuned version of bert-base-uncased on the Microsoft Research Paraphrase Corpus (MRPC) dataset for paraphrase detection using the MRPC dataset.

Intended uses & limitations

This model is intended for paraphrase detection tasks, particularly those similar to the MRPC dataset. It may not perform well on substantially different datasets or tasks.

Training and evaluation data

The model was trained on the MRPC dataset, which contains 5,801 sentence pairs extracted from news sources on the web. 3,900 pairs were labeled as paraphrases by human annotators.

Training procedure

The model was fine-tuned using the Hugging Face Transformers library. We used a batch size of 16, learning rate of 2e-5, and trained for 3 epochs.

Evaluation results

The model achieved the following results on the MRPC validation set:

  • Accuracy: 0.8480
  • F1 Score: 0.8927