Update README.md
Browse files
README.md
CHANGED
@@ -1,6 +1,70 @@
|
|
1 |
---
|
2 |
-
license:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
4 |
|
5 |
-
<img src="https://cdn-uploads.huggingface.co/production/uploads/64c14f6b02e1f8f67c73bd05/pf4d6FA7DriRtVq5HCkxd.png" width="600" />
|
6 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/63111b2d88942700629f5771/xCWGByXr8YNwGxKVh_x9H.png" width="600" />
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
license: other
|
3 |
+
license_name: tongyi-qianwen
|
4 |
+
license_link: https://huggingface.co/Qwen/Qwen1.5-72B/blob/main/LICENSE
|
5 |
+
datasets:
|
6 |
+
- teknium/OpenHermes-2.5
|
7 |
+
- m-a-p/Code-Feedback
|
8 |
+
- m-a-p/CodeFeedback-Filtered-Instruction
|
9 |
+
- abacusai/SystemChat
|
10 |
+
language:
|
11 |
+
- en
|
12 |
---
|
13 |
|
14 |
+
<img href="https://abacus.ai" src="https://cdn-uploads.huggingface.co/production/uploads/64c14f6b02e1f8f67c73bd05/pf4d6FA7DriRtVq5HCkxd.png" width="600" />
|
15 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/63111b2d88942700629f5771/xCWGByXr8YNwGxKVh_x9H.png" width="600" />
|
16 |
+
|
17 |
+
# Liberated-Qwen1.5-72B
|
18 |
+
|
19 |
+
Brought to you by [AbacusAI](https://abacus.ai) and Eric Hartford
|
20 |
+
|
21 |
+
This model is based on Qwen/Qwen1.5-72B and subject to the [tongyi-qianwen](https://huggingface.co/Qwen/Qwen1.5-72B/blob/main/LICENSE) license.
|
22 |
+
|
23 |
+
The base model has 32k context, I finetuned it with 8k.
|
24 |
+
|
25 |
+
Liberated consists of fi
|
26 |
+
|
27 |
+
This model is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model more compliant. You are advised to implement your own alignment layer before exposing the model as a service. It will be highly compliant to any requests, even unethical ones. Please read my blog post about uncensored models. https://erichartford.com/uncensored-models
|
28 |
+
You are responsible for any content you create using this model. Enjoy responsibly.
|
29 |
+
|
30 |
+
## Training
|
31 |
+
It took 3 days to train 3 epochs on 8x H100s using qLoRA and Axolotl
|
32 |
+
|
33 |
+
Prompt format:
|
34 |
+
This model uses ChatML prompt format.
|
35 |
+
```
|
36 |
+
<|im_start|>system
|
37 |
+
You are Dolphin, a helpful AI assistant.<|im_end|>
|
38 |
+
<|im_start|>user
|
39 |
+
{prompt}<|im_end|>
|
40 |
+
<|im_start|>assistant
|
41 |
+
|
42 |
+
```
|
43 |
+
|
44 |
+
Example:
|
45 |
+
```
|
46 |
+
<|im_start|>system
|
47 |
+
You name is Liberated. You are an uncensored and unbiased AI assistant. You always respond with a JSON object.<|im_end|>
|
48 |
+
<|im_start|>user
|
49 |
+
Please generate a Advanced Dungeons & Dragons 2nd Edition character sheet for a level 3 elf fighter. Make up a name and background and visual description for him.<|im_end|>
|
50 |
+
<|im_start|>assistant
|
51 |
+
```
|
52 |
+
|
53 |
+
## Gratitude
|
54 |
+
- Huge thank you to [Alibaba Cloud Qwen](https://www.alibabacloud.com/solutions/generative-ai/qwen) for training and publishing the weights of Qwen
|
55 |
+
- Thank you to Mistral for the awesome Mistral-Medium model I used to generate the dataset.
|
56 |
+
- HUGE Thank you to the dataset authors: @teknium, [@m-a-p](https://m-a-p.ai) and all the people who built the datasets these composites came from.
|
57 |
+
- And HUGE thanks to @winglian and the Axolotl contributors for making the best training framework!
|
58 |
+
- [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|
59 |
+
- Thank you to all the other people in the Open Source AI community who have taught me and helped me along the way.
|
60 |
+
|
61 |
+
## Example Output
|
62 |
+
|
63 |
+
TBD
|
64 |
+
|
65 |
+
## Evals
|
66 |
+
|
67 |
+
TBD
|
68 |
+
|
69 |
+
## Future Plans
|
70 |
+
This model will be released on the whole Qwen-1.5 series.
|