AMDBartek commited on
Commit
d3f49d2
1 Parent(s): ef18749

Mention and thank other quants, tidy README

Browse files
Files changed (1) hide show
  1. README.md +7 -4
README.md CHANGED
@@ -78,7 +78,7 @@ A quick overview of the model's strengths include:
78
  - Strong system prompt adherence
79
  - Best-in-class multilingual capabilities compared to competing models of its size (English, Chinese, Spanish, French, Polish, and more!)
80
  - Being unbiased and truthful (although, you should note that all forms of intelligence can and WILL make mistakes, whether organic or artificial)
81
- - Having no unecessary censorship (some unfortunately bleeds through since `Meta-Llama-3-8B-Instruct` was used as a base and version 3 should fix that)
82
  - Simply being fun to talk to
83
 
84
  ## Model Specs
@@ -94,13 +94,16 @@ A quick overview of the model's strengths include:
94
  Well, yes, but actually no. You may see the names of benchmarks in the datasets used, however only **train** splits were used. If you don't know the difference, please learn.
95
 
96
  ## Quants and Other Formats
97
- - GGUFs: [https://huggingface.co/darkcloudai/huskylm-2.5-8b-GGUF](https://huggingface.co/darkcloudai/huskylm-2.5-8b-GGUF)
98
- - AWQ (bits: 4, gs: 128, version: gemm): [https://huggingface.co/darkcloudai/huskylm-2.5-8b-AWQ](https://huggingface.co/darkcloudai/huskylm-2.5-8b-AWQ)
 
 
 
99
 
100
  ## Huge Thank You to the Following People/Companies
101
 
102
  - [Meta AI](https://llama.meta.com/llama3/): This model would never have been possible if Meta AI did not release Llama 3 with an open license. We thank them deeply for making frontier LLMs available for all.
103
- - [Jon Durbin](https://huggingface.co/jondurbin): We've used many of his datasets to train this model, specifically `airoboros-3.2`, `contextual-dpo-v0.1`, `gutenberg-dpo-v0.1`, `py-dpo-v0.1`, `truthy-dpo-v0.1`, `cinematika-v0.1`, `gutenberg-dpo-v0.1`. His work is amazing and we thank him a lot. We've used a lot of datasets for our model that he used for his `bagel` series of models too. If you couldn't already guess, this model is essentially a `bagel` model but with our custom datasets and RLAIF methodology added in.
104
  - [Hugging Face](https://github.com/huggingface): Throughout Darkcloud AI's life, we've extensively used and relied on libraries made by HuggingFace and we thank them and everyone who has contributed.
105
  - [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl): We've used Axolotl to streamline the (SFT) fine-tuning of our LLMs. Huge thank you to them and every contributor.
106
  - You: That's right! You, the user. We value every single bit of feedback we receive from our users as it helps us to make our models better for everyone. If you have any issues, *please* give feedback. Every little bit of information helps, no matter how minor the issue or question you have is!
 
78
  - Strong system prompt adherence
79
  - Best-in-class multilingual capabilities compared to competing models of its size (English, Chinese, Spanish, French, Polish, and more!)
80
  - Being unbiased and truthful (although, you should note that all forms of intelligence can and WILL make mistakes, whether organic or artificial)
81
+ - Having no unecessary censorship (some unfortunately bleeds through since `Meta-Llama-3-8B-Instruct` was used as a base and HuskyLM 3 should fix that - we're training from the ground up from the base Llama 3 next time)
82
  - Simply being fun to talk to
83
 
84
  ## Model Specs
 
94
  Well, yes, but actually no. You may see the names of benchmarks in the datasets used, however only **train** splits were used. If you don't know the difference, please learn.
95
 
96
  ## Quants and Other Formats
97
+ - GGUFs:
98
+ * Official: [https://huggingface.co/darkcloudai/huskylm-2.5-8b-GGUF](https://huggingface.co/darkcloudai/huskylm-2.5-8b-GGUF)
99
+ * mradermacher's static quants (thank you!): [https://huggingface.co/mradermacher/huskylm-2.5-8b-GGUF](https://huggingface.co/mradermacher/huskylm-2.5-8b-GGUF)
100
+ * mradermacher's imatrix quants (thank you!): [https://huggingface.co/mradermacher/huskylm-2.5-8b-i1-GGUF](https://huggingface.co/mradermacher/huskylm-2.5-8b-i1-GGUF)
101
+ - Official AWQ (bits: 4, gs: 128, version: gemm): [https://huggingface.co/darkcloudai/huskylm-2.5-8b-AWQ](https://huggingface.co/darkcloudai/huskylm-2.5-8b-AWQ)
102
 
103
  ## Huge Thank You to the Following People/Companies
104
 
105
  - [Meta AI](https://llama.meta.com/llama3/): This model would never have been possible if Meta AI did not release Llama 3 with an open license. We thank them deeply for making frontier LLMs available for all.
106
+ - [Jon Durbin](https://huggingface.co/jondurbin): We've used many of his datasets to train this model, specifically `airoboros-3.2`, `contextual-dpo-v0.1`, `gutenberg-dpo-v0.1`, `py-dpo-v0.1`, `truthy-dpo-v0.1`, `cinematika-v0.1`, `gutenberg-dpo-v0.1`. His work is amazing and we thank him a lot. We've used a lot of datasets for our model that he used for his `bagel` series of models too. If you couldn't already guess, this model is essentially a `bagel`-type model but with our custom datasets and RLAIF methodology added in.
107
  - [Hugging Face](https://github.com/huggingface): Throughout Darkcloud AI's life, we've extensively used and relied on libraries made by HuggingFace and we thank them and everyone who has contributed.
108
  - [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl): We've used Axolotl to streamline the (SFT) fine-tuning of our LLMs. Huge thank you to them and every contributor.
109
  - You: That's right! You, the user. We value every single bit of feedback we receive from our users as it helps us to make our models better for everyone. If you have any issues, *please* give feedback. Every little bit of information helps, no matter how minor the issue or question you have is!