922CA commited on
Commit
6f42368
·
1 Parent(s): 8721061

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -0
README.md ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama2
3
+ ---
4
+ # Taga-llama-v0.3a:
5
+ * Test Tagalog model, particularly with using a partially synthetic dataset (observing the Tagalog capability of LLaMA-2)
6
+ * Fine tuned on dataset of ~1k items (Tagalog/Taglish dataset, based off Tagalog sentences augmented by LLaMA-2-13b base to create a 3-turn dialogue dataset between Human and Assistant); dataset to be further refined
7
+ * Trained on llama 7b chat
8
+ * May still switch to Taglish or English: please see usage
9
+ * [QLoras (hf and GGML)](https://huggingface.co/922-Narra/tagalog-lm-lora-tests/tree/main/llama-2-7b-chat-tagalog-0.3a)
10
+
11
+ ### USAGE
12
+ This is meant to be mainly a chat model.
13
+
14
+ Use "Human" and "Assistant" and prompt with Tagalog. Example:
15
+
16
+ "Ito ay isang chat log sa pagitan ng AI Assistant na nagta-Tagalog at isang Pilipino. Magsimula ng chat:\nHuman: Hello po?\nAssistant:"
17
+
18
+ ### HYPERPARAMS
19
+ * Trained for 2 epochs
20
+ * rank: 16
21
+ * lora alpha: 32
22
+ * lora dropout: 0.5
23
+ * lr: 2e-4
24
+ * batch size: 2
25
+ * warmup ratio: 0.075
26
+ * grad steps: 4
27
+
28
+ ### WARNINGS AND DISCLAIMERS
29
+ Note that aside from formatting and other minor edits, dataset used is mostly as is generated by LM. As such, while this version may be better at coherency or chatting than previous Tagalog ones, conversations may still switch between languages or easily derail.
30
+
31
+ Finally, this model is not guaranteed to output aligned or safe outputs nor is it meant for production use - use at your own risk!