Neko-Institute-of-Science
commited on
Commit
•
576e8c6
1
Parent(s):
9574884
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
datasets:
|
3 |
+
- gozfarb/ShareGPT_Vicuna_unfiltered
|
4 |
+
---
|
5 |
+
# Convert tools
|
6 |
+
https://github.com/practicaldreamer/vicuna_to_alpaca
|
7 |
+
|
8 |
+
# Training tool
|
9 |
+
https://github.com/oobabooga/text-generation-webui
|
10 |
+
|
11 |
+
ATM I'm using 2023.05.04v0 of the dataset and training full context.
|
12 |
+
|
13 |
+
# How to test?
|
14 |
+
1. Download LLaMA-13B-HF: https://huggingface.co/Neko-Institute-of-Science/LLaMA-30B-HF
|
15 |
+
2. Replace special_tokens_map.json and tokenizer_config.json using the ones on this repo.
|
16 |
+
3. Rename LLaMA-30B-HF to vicuna-30b
|
17 |
+
4. Load ooba: ```python server.py --listen --model vicuna-30b --load-in-8bit --chat --lora checkpoint-xxxx```
|
18 |
+
5. Instruct mode: Vicuna-v1 it will load Vicuna-v0 by defualt
|
19 |
+
|
20 |
+
|
21 |
+
# Want to see it Training?
|
22 |
+
https://wandb.ai/neko-science/VicUnLocked/runs/vx8yzwi7
|