Update README.md
Browse files
README.md
CHANGED
@@ -26,9 +26,9 @@ pipeline_tag: text-generation
|
|
26 |
</div>
|
27 |
<!-- header end -->
|
28 |
|
29 |
-
#
|
30 |
- Model creator: [Stability AI](https://huggingface.co/stabilityai)
|
31 |
-
- Original model: [
|
32 |
|
33 |
## Description
|
34 |
|
@@ -40,8 +40,9 @@ None
|
|
40 |
|
41 |
## Repositories available
|
42 |
|
43 |
-
* [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/
|
44 |
-
* [
|
|
|
45 |
|
46 |
## Prompt template: Orca-Hashes
|
47 |
|
@@ -211,25 +212,25 @@ Thank you to all my generous patrons and donaters!
|
|
211 |
|
212 |
<!-- footer end -->
|
213 |
|
214 |
-
# Original model card: Stability AI's
|
215 |
|
216 |
-
#
|
217 |
|
218 |
## Model Description
|
219 |
|
220 |
-
`
|
221 |
|
222 |
## Usage
|
223 |
|
224 |
-
Start chatting with `
|
225 |
|
226 |
```python
|
227 |
import torch
|
228 |
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
229 |
|
230 |
-
tokenizer = AutoTokenizer.from_pretrained("stabilityai/
|
231 |
-
model = AutoModelForCausalLM.from_pretrained("stabilityai/
|
232 |
-
system_prompt = "### System:\nYou are
|
233 |
|
234 |
message = "Write me a poem please"
|
235 |
prompt = f"{system_prompt}### User: {message}\n\n### Assistant:\n"
|
@@ -239,7 +240,7 @@ output = model.generate(**inputs, do_sample=True, top_p=0.95, top_k=0, max_new_t
|
|
239 |
print(tokenizer.decode(output[0], skip_special_tokens=True))
|
240 |
```
|
241 |
|
242 |
-
|
243 |
```
|
244 |
### System:
|
245 |
This is a system prompt, please behave and help the user.
|
@@ -247,22 +248,22 @@ This is a system prompt, please behave and help the user.
|
|
247 |
### User:
|
248 |
Your prompt here
|
249 |
|
250 |
-
### Assistant
|
251 |
-
The output of
|
252 |
```
|
253 |
|
254 |
## Model Details
|
255 |
|
256 |
* **Developed by**: [Stability AI](https://stability.ai/)
|
257 |
-
* **Model type**:
|
258 |
* **Language(s)**: English
|
259 |
* **Library**: [HuggingFace Transformers](https://github.com/huggingface/transformers)
|
260 |
-
* **License**: Fine-tuned checkpoints (`
|
261 |
* **Contact**: For questions and comments about the model, please email `[email protected]`
|
262 |
|
263 |
### Training Dataset
|
264 |
|
265 |
-
`
|
266 |
|
267 |
### Training Procedure
|
268 |
|
@@ -273,15 +274,9 @@ Models are learned via supervised fine-tuning on the aforementioned datasets, tr
|
|
273 |
| Orca pt1 packed | 256 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
|
274 |
| Orca pt2 unpacked | 512 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
|
275 |
|
276 |
-
##
|
277 |
-
|
278 |
-
### Intended Use
|
279 |
-
|
280 |
-
These models are intended for research only, in adherence with the [CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/) license.
|
281 |
|
282 |
-
|
283 |
-
|
284 |
-
Although the aforementioned dataset helps to steer the base language models into "safer" distributions of text, not all biases and toxicity can be mitigated through fine-tuning. We ask that users be mindful of such potential issues that can arise in generated responses. Do not treat model outputs as substitutes for human judgment or as sources of truth. Please use it responsibly.
|
285 |
|
286 |
## Citations
|
287 |
|
@@ -305,4 +300,4 @@ Although the aforementioned dataset helps to steer the base language models into
|
|
305 |
archivePrefix={arXiv},
|
306 |
primaryClass={cs.CL}
|
307 |
}
|
308 |
-
```
|
|
|
26 |
</div>
|
27 |
<!-- header end -->
|
28 |
|
29 |
+
# StableBeluga 2 - GGML
|
30 |
- Model creator: [Stability AI](https://huggingface.co/stabilityai)
|
31 |
+
- Original model: [StableBeluga 2](https://huggingface.co/stabilityai/StableBeluga2)
|
32 |
|
33 |
## Description
|
34 |
|
|
|
40 |
|
41 |
## Repositories available
|
42 |
|
43 |
+
* [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/StableBeluga2-GPTQ)
|
44 |
+
* [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/StableBeluga2-GGML)
|
45 |
+
* [Stability AI's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/stabilityai/StableBeluga2)
|
46 |
|
47 |
## Prompt template: Orca-Hashes
|
48 |
|
|
|
212 |
|
213 |
<!-- footer end -->
|
214 |
|
215 |
+
# Original model card: Stability AI's StableBeluga 2
|
216 |
|
217 |
+
# Stable Beluga 2
|
218 |
|
219 |
## Model Description
|
220 |
|
221 |
+
`Stable Beluga 2` is a Llama2 70B model finetuned on an Orca style Dataset
|
222 |
|
223 |
## Usage
|
224 |
|
225 |
+
Start chatting with `Stable Beluga 2` using the following code snippet:
|
226 |
|
227 |
```python
|
228 |
import torch
|
229 |
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
230 |
|
231 |
+
tokenizer = AutoTokenizer.from_pretrained("stabilityai/StableBeluga2", use_fast=False)
|
232 |
+
model = AutoModelForCausalLM.from_pretrained("stabilityai/StableBeluga2", torch_dtype=torch.float16, low_cpu_mem_usage=True, device_map="auto")
|
233 |
+
system_prompt = "### System:\nYou are Stable Beluga, an AI that follows instructions extremely well. Help as much as you can. Remember, be safe, and don't do anything illegal.\n\n"
|
234 |
|
235 |
message = "Write me a poem please"
|
236 |
prompt = f"{system_prompt}### User: {message}\n\n### Assistant:\n"
|
|
|
240 |
print(tokenizer.decode(output[0], skip_special_tokens=True))
|
241 |
```
|
242 |
|
243 |
+
Stable Beluga 2 should be used with this prompt format:
|
244 |
```
|
245 |
### System:
|
246 |
This is a system prompt, please behave and help the user.
|
|
|
248 |
### User:
|
249 |
Your prompt here
|
250 |
|
251 |
+
### Assistant:
|
252 |
+
The output of Stable Beluga 2
|
253 |
```
|
254 |
|
255 |
## Model Details
|
256 |
|
257 |
* **Developed by**: [Stability AI](https://stability.ai/)
|
258 |
+
* **Model type**: Stable Beluga 2 is an auto-regressive language model fine-tuned on Llama2 70B.
|
259 |
* **Language(s)**: English
|
260 |
* **Library**: [HuggingFace Transformers](https://github.com/huggingface/transformers)
|
261 |
+
* **License**: Fine-tuned checkpoints (`Stable Beluga 2`) is licensed under the [STABLE BELUGA NON-COMMERCIAL COMMUNITY LICENSE AGREEMENT](https://huggingface.co/stabilityai/StableBeluga2/blob/main/LICENSE.txt)
|
262 |
* **Contact**: For questions and comments about the model, please email `[email protected]`
|
263 |
|
264 |
### Training Dataset
|
265 |
|
266 |
+
` Stable Beluga 2` is trained on our internal Orca-style dataset
|
267 |
|
268 |
### Training Procedure
|
269 |
|
|
|
274 |
| Orca pt1 packed | 256 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
|
275 |
| Orca pt2 unpacked | 512 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
|
276 |
|
277 |
+
## Ethical Considerations and Limitations
|
|
|
|
|
|
|
|
|
278 |
|
279 |
+
Beluga is a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Beluga's potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses to user prompts. Therefore, before deploying any applications of Beluga, developers should perform safety testing and tuning tailored to their specific applications of the model.
|
|
|
|
|
280 |
|
281 |
## Citations
|
282 |
|
|
|
300 |
archivePrefix={arXiv},
|
301 |
primaryClass={cs.CL}
|
302 |
}
|
303 |
+
```
|