metadata
language: en
tags:
- bert
- sequence-classification
- mrpc
- paraphrase
license: mit
Model description
Fine-tuned version of bert-base-uncased
on the Microsoft Research Paraphrase Corpus (MRPC) dataset for paraphrase detection using the MRPC dataset.
Intended uses & limitations
This model is intended for paraphrase detection tasks, particularly those similar to the MRPC dataset. It may not perform well on substantially different datasets or tasks.
Training and evaluation data
The model was trained on the MRPC dataset, which contains 5,801 sentence pairs extracted from news sources on the web. 3,900 pairs were labeled as paraphrases by human annotators.
Training procedure
The model was fine-tuned using the Hugging Face Transformers library. We used a batch size of 16, learning rate of 2e-5, and trained for 3 epochs.
Evaluation results
The model achieved the following results on the MRPC validation set:
- Accuracy: 0.8480
- F1 Score: 0.8927