Angainor commited on
Commit
6b7dd5a
1 Parent(s): e50db28

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -1
README.md CHANGED
@@ -1,3 +1,10 @@
 
 
 
 
 
 
 
1
  This repo contains a low-rank adapter for LLaMA-13b fit on the Stanford Alpaca dataset.
2
 
3
  This version of the weights was trained on dual RTX3090 with the following hyperparameters:
@@ -21,4 +28,8 @@ OMP_NUM_THREADS=4 WORLD_SIZE=2 CUDA_VISIBLE_DEVICES=0,1 torchrun --nproc_per_nod
21
  --val_set_size=0 \
22
  --micro_batch_size=32
23
 
24
- Instructions for running it can be found at https://github.com/tloen/alpaca-lora.
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - yahma/alpaca-cleaned
5
+ language:
6
+ - en
7
+ ---
8
  This repo contains a low-rank adapter for LLaMA-13b fit on the Stanford Alpaca dataset.
9
 
10
  This version of the weights was trained on dual RTX3090 with the following hyperparameters:
 
28
  --val_set_size=0 \
29
  --micro_batch_size=32
30
 
31
+ LR warmup was tuned to fit the first epoch.
32
+
33
+ Instructions for running it can be found at https://github.com/tloen/alpaca-lora.
34
+
35
+ ![10 epochs](alpaca13b.png)