File size: 809 Bytes
b82cc5b ecc1972 ce5973a b82cc5b ce5973a b82cc5b 5e18787 b82cc5b ecc1972 b82cc5b 5e18787 b82cc5b 5e18787 b82cc5b 5e18787 b82cc5b 5e18787 b82cc5b 5e18787 b82cc5b 5e18787 b82cc5b ee3da99 b82cc5b ee3da99 b82cc5b ee3da99 b82cc5b ee3da99 b82cc5b |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 |
---
base_model: google/gemma-2b
datasets:
- tatsu-lab/alpaca
language: en
tags:
- torchtune
---
# My Torchtune Model
This model is a finetuned version of [google/gemma-2b](https://huggingface.co/google/gemma-2b)
# Model description
More information needed
# Training and evaluation results
More information needed
# Training procedure
This model was trained using the [torchtune](https://github.com/pytorch/torchtune) library using the following command:
```bash
/Users/salmanmohammadi/projects/torchtune/recipes/lora_finetune_single_device.py --config /Users/salmanmohammadi/projects/torchtune/recipes/configs/gemma/2B_lora_single_device.yaml \
device=mps \
epochs=1 \
max_steps_per_epoch=10
```
# Framework versions
- torchtune 0.0.0
- torchao 0.5.0
- datasets 2.20.0
- sentencepiece 0.2.0
|