--- language: - en - ms - zh tags: - sentiment-analysis - text-classification - multilingual license: apache-2.0 datasets: - tyqiangz/multilingual-sentiments metrics: - accuracy model-index: - name: xlm-roberta-base-sentiment-multilingual-finetuned results: - task: type: text-classification name: Text Classification dataset: type: tyqiangz/multilingual-sentiments name: Multilingual Sentiments metrics: - type: accuracy value: 0.7737435897435897 # Replace with your model's accuracy --- # xlm-roberta-base-sentiment-multilingual-finetuned ## Model description This is a fine-tuned version of the [cardiffnlp/twitter-xlm-roberta-base-sentiment-multilingual](https://huggingface.co/cardiffnlp/twitter-xlm-roberta-base-sentiment-multilingual) model, trained on the [tyqiangz/multilingual-sentiments](https://huggingface.co/datasets/tyqiangz/multilingual-sentiments) dataset. It's designed for multilingual sentiment analysis in English, Malay, and Chinese. ## Intended uses & limitations This model is intended for sentiment analysis tasks in English, Malay, and Chinese. It can classify text into three sentiment categories: positive, negative, and neutral. ## Training and evaluation data The model was trained and evaluated on the [tyqiangz/multilingual-sentiments](https://huggingface.co/datasets/tyqiangz/multilingual-sentiments) dataset, which includes data in English, Malay, and Chinese. ## Training procedure The model was fine-tuned using the Hugging Face Transformers library. training_args = TrainingArguments( output_dir="./results", num_train_epochs=1, per_device_train_batch_size=16, per_device_eval_batch_size=64, warmup_steps=500, weight_decay=0.01, logging_dir='./logs', logging_steps=10, evaluation_strategy="epoch", save_strategy="epoch", load_best_model_at_end=True, ) ## Evaluation results 'eval_accuracy': 0.7737435897435897, 'eval_f1': 0.7724731131078052, 'eval_precision': 0.7733524717389839, 'eval_recall': 0.7737435897435897, ## Environmental impact Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).