metadata
license: apache-2.0
library_name: peft
tags:
- trl
- sft
- generated_from_trainer
base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
model-index:
- name: tinyllama-tarot-v1
results: []
pipeline_tag: text2text-generation
widget:
- text: >-
Give me a one paragraph tarot reading if I pull the cards Nine of Cups,
King of Pentacles and Three of Cups.
datasets:
- barissglc/tarot
tinyllama-tarot-v1
This model is a fine-tuned version of TinyLlama/TinyLlama-1.1B-Chat-v1.0 .
Model description
This model is a language model capable of making predictions based on tarot cards. Trained to respond to questions related to topics such as love, career, and general life, tarot cards are the foundation of its predictions. The model can make predictions based on the selected tarot cards. You can access the tarot cards from the tarot dataset.
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- training_steps: 250
- mixed_precision_training: Native AMP
Training results
Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.1
- Tokenizers 0.15.2