|
Dummy Model for Lab4 |
|
|
|
This model is a fine-tuned version of bert-base-uncased on SST-2 dataset. |
|
|
|
Results of the evaluation set: |
|
|
|
Accuracy: 0.64 |
|
|
|
This model was fine-tuneded for personal research usage. |
|
with randomly selected 100 training datas and 100 evaluation datas from SST-2 dataset. |
|
|
|
|
|
# Evaluation |
|
import evaluate |
|
predictions = trainer.predict(Resrt_eval) |
|
print(predictions.predictions.shape, predictions.label_ids.shape) |
|
preds = np.argmax(predictions.predictions, axis=-1) |
|
|
|
|
|
metric = evaluate.load("glue", "sst2") |
|
metric.compute(predictions=preds, references=predictions.label_ids) |
|
|
|
Training hyperparameters |
|
The following hyperparameters were used during training: |
|
|
|
learning_rate: unset |
|
train_batch_size: unset |
|
eval_batch_size: unset |
|
seed of training dataset: 49282927487 |
|
seed of evaluation dataset:492829487 |
|
|
|
lr_scheduler_type: linear |
|
num_epochs: 3.0 |
|
Training results |
|
|
|
Epoch Training Loss Validation Loss |
|
1 No log 0.674658 0.480000 |
|
2 No log 0.640980 0.600000 |
|
3 No log 0.640266 0.640000 |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65e812f328ebe0129dd9a2b4/mxsW8uXzCJrVbmamFFfqy.png) |
|
|
|
|