mental's picture
Update README.md (#1)
29ad3d2
metadata
license: cc-by-nc-4.0

MentalBERT is a model initialized with RoBERTa-large (uncased_L-24_H-1024_A-16) and trained with mental health-related posts collected from Reddit.

We follow the standard pretraining protocols of BERT and RoBERTa with Huggingface’s Transformers library.

We use four Nvidia Tesla v100 GPUs to train the two language models. We set the batch size to 8 per GPU, evaluate every 1,000 steps, and train for 312,000 iterations.

Usage

Load the model via Huggingface’s Transformers library:

from transformers import AutoTokenizer, AutoModel
tokenizer = AutoTokenizer.from_pretrained("AIMH/mental-roberta-large")
model = AutoModel.from_pretrained("AIMH/mental-roberta-large")

To minimize the influence of worrying mask predictions, this model is gated. To download a gated model, you’ll need to be authenticated. Know more about gated models.

Paper

MentalBERT: Publicly Available Pretrained Language Models for Mental Healthcare.

@inproceedings{ji2022mentalbert,
  title     = {{MentalBERT: Publicly Available Pretrained Language Models for Mental Healthcare}},
  author    = {Shaoxiong Ji and Tianlin Zhang and Luna Ansari and Jie Fu and Prayag Tiwari and Erik Cambria},
  year      = {2022},
  booktitle = {Proceedings of LREC}
}