Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
This repo contains a low-rank adapter for LLaMA-13b fit on the Stanford Alpaca dataset.
|
2 |
|
3 |
This version of the weights was trained on dual RTX3090 with the following hyperparameters:
|
@@ -21,4 +28,8 @@ OMP_NUM_THREADS=4 WORLD_SIZE=2 CUDA_VISIBLE_DEVICES=0,1 torchrun --nproc_per_nod
|
|
21 |
--val_set_size=0 \
|
22 |
--micro_batch_size=32
|
23 |
|
24 |
-
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
datasets:
|
4 |
+
- yahma/alpaca-cleaned
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
---
|
8 |
This repo contains a low-rank adapter for LLaMA-13b fit on the Stanford Alpaca dataset.
|
9 |
|
10 |
This version of the weights was trained on dual RTX3090 with the following hyperparameters:
|
|
|
28 |
--val_set_size=0 \
|
29 |
--micro_batch_size=32
|
30 |
|
31 |
+
LR warmup was tuned to fit the first epoch.
|
32 |
+
|
33 |
+
Instructions for running it can be found at https://github.com/tloen/alpaca-lora.
|
34 |
+
|
35 |
+
![10 epochs](alpaca13b.png)
|