BLOOMChat V2.0
BLOOMChat-v2 is a 176 billion parameter multilingual chat model. It is finetuned from BLOOM (176B) on long-sequence multilingual data and assistant-style conversation datasets. It supports conversation, question answering and generative answers in multiple languages.
Model Details
Model Description
- Developed by: SambaNova Systems
- Model type: Language Model
- Language(s): Multiple; see training data from BLOOM
- License: BLOOMChat-176B LICENSE v1.0
- Finetuned from model: BigScience Group BLOOM
Basic Information
Licensing
To increase accessibility and to support the open-source community, SambaNova is releasing BLOOMChat under a modified version of the Apache 2.0 license, which includes use-based restrictions from BLOOM’s RAIL license. While use-based restrictions are necessarily passed through, there are no blanket restrictions on reuse, distribution, commercialization or adaptation. Please review SambaNova’s BLOOMChat-176B License
Uses
Click to expand
Direct Use
This model is intended for commercial and research use.
Out-of-Scope Use
BLOOMChat should NOT be used for:
- Mission-critical applications
- Applications that involve the safety of others
- Making highly important decisions
- Important automated pipelines
This model is still in early development and can be prone to mistakes and hallucinations, there is still room for improvement. This model is intended to provide the community with a multilingual chat LLM baseline.
Recommendations
Users should be made aware of the risks, biases, limitations, and restrictions of the model, which are listed down at the bottom of the page.
How to Get Started with the Model
Click to expand
Loading in model with Huggingface
from transformers import AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("sambanovasystems/BLOOMChat-176B-v2")
model = AutoModelForCausalLM.from_pretrained("sambanovasystems/BLOOMChat-176B-v2", device_map="auto", torch_dtype="auto")
Quick Start Inference on SambaNova's in-house Reconfigurable Dataflow Unit (RDU)
The inference code to run the model can be found our github repo. This code requires the SambaFlow SDK to execute. For those interested in running models on RDUs, please feel free to get in touch.
Quick Start Inference on GPU
First create a python virtual environment for these packages
python3 -m venv bloomchat_venv
source bloomchat_venv/bin/activate
pip install --upgrade pip
pip install flask flask_api gunicorn pydantic accelerate huggingface_hub>=0.9.0 deepspeed>=0.7.3 deepspeed-mii==0.0.2
And then
pip install transformers==4.27.0
You will see messages like this
ERROR: deepspeed-mii 0.0.2 has requirement transformers==4.21.2, but you'll have transformers 4.27.0 which is incompatible.
Installing collected packages: transformers
Found existing installation: transformers 4.21.2
Uninstalling transformers-4.21.2:
Successfully uninstalled transformers-4.21.2
Successfully installed transformers-4.27.0
Now let's git clone the huggingface/transformers-bloom-inference repo.
git clone https://github.com/huggingface/transformers-bloom-inference.git
cd transformers-bloom-inference/
And then you need to modify two files in this transformers-bloom-inference repo:
- Modifying
inference_server/models/hf_accelerate.py
- This is because for our testing of this repo we used 4 80GB A100 GPUs and would run into memory issues
- Modifying
inference_server/cli.py
- This is because the model was trained using specific human, bot tags
- Trailing spaces may lead to subpar performance
Modifications for inference_server/models/hf_accelerate.py
:
diff --git a/inference_server/models/hf_accelerate.py b/inference_server/models/hf_accelerate.py
index 9be3c3f..a8ecb1d 100644
--- a/inference_server/models/hf_accelerate.py
+++ b/inference_server/models/hf_accelerate.py
@@ -1,4 +1,5 @@
from argparse import Namespace
+from accelerate.utils.modeling import get_max_memory
import torch
@@ -12,6 +13,12 @@ class HFAccelerateModel(Model):
kwargs = {"pretrained_model_name_or_path": args.model_name, "device_map": "auto"}
+ original_max_memory_dict = get_max_memory()
+
+ reduce_max_memory_dict = {device_key: int(original_max_memory_dict[device_key] * 0.85) for device_key in original_max_memory_dict}
+
+ kwargs["max_memory"] = reduce_max_memory_dict
+
if get_world_size() > 1:
kwargs["device_map"] = "balanced_low_0"
Modifications for inference_server/cli.py
:
diff --git a/inference_server/cli.py b/inference_server/cli.py
index fc903d5..5450236 100644
--- a/inference_server/cli.py
+++ b/inference_server/cli.py
@@ -22,6 +22,9 @@ def main() -> None:
while True:
input_text = input("Input text: ")
+ input_text = input_text.strip()
+ modified_input_text = f"<human>: {input_text}\n<bot>:"
+
if input("change generate_kwargs? [y/n] ") == "y":
while True:
try:
@@ -33,7 +36,7 @@ def main() -> None:
print("message =", e_message)
continue
- response = model.generate(text=[input_text], generate_kwargs=generate_kwargs)
+ response = model.generate(text=[modified_input_text], generate_kwargs=generate_kwargs)
print_rank_0("Output text:", response.text[0])
print_rank_0("Generated tokens:", response.num_generated_tokens[0])
And now you are good to go!
Running command for bf16, NO sampling
python -m inference_server.cli --model_name sambanovasystems/BLOOMChat-176B-v2 --model_class AutoModelForCausalLM --dtype bf16 --deployment_framework hf_accelerate --generate_kwargs '{"do_sample": false, "max_new_tokens": 512}'
Running command for bf16, YES sampling
python -m inference_server.cli --model_name sambanovasystems/BLOOMChat-176B-v2 --model_class AutoModelForCausalLM --dtype bf16 --deployment_framework hf_accelerate --generate_kwargs '{"do_sample": true, "temperature": 0.8, "repetition_penalty": 1.2, "top_p": 0.9, "max_new_tokens": 512}'
Running command for int8 (sub optimal performance, but fast inference time) NO sampling:
python -m inference_server.cli --model_name sambanovasystems/BLOOMChat-176B-v2 --model_class AutoModelForCausalLM --dtype int8 --deployment_framework hf_accelerate --generate_kwargs '{"do_sample": false, "max_new_tokens": 512}'
Running command for int8 (sub optimal performance, but fast inference time) YES sampling:
python -m inference_server.cli --model_name sambanovasystems/BLOOMChat-176B-v2 --model_class AutoModelForCausalLM --dtype int8 --deployment_framework hf_accelerate --generate_kwargs '{"do_sample": true, "temperature": 0.8, "repetition_penalty": 1.2, "top_p": 0.9, "max_new_tokens": 512}'
DISCLAIMER: When using int8, the results will be subpar compared to bf16 as the model is being quantized.
Suggested Inference Parameters
- Temperature: 0.8
- Repetition penalty: 1.2
- Top-p: 0.9
- Max generated tokens: 512
Suggested Prompts To Try in GPU Tutorial
Input text: Write a script in which Bob accidentally breaks his dad's guitar
Input text: Create an itemized list of tasks to complete to start a clothing brand
Input text: 十七岁的风是什么颜色的?
Some example completions for English
Click to expand
<human>: What does it mean for the Fed to raise or lower interest rates?
<bot>: Lowering interest rates means that banks will be able to borrow from the Federal Reserve at a cheaper rate. This makes it more attractive for banks to borrow money and increase their lending activities, as lowering interest rates makes it more attractive for people to borrow money. This increases the money supply and is typically seen as a stimulus for the economy. Conversely, raising interest rates makes it more expensive for banks to borrow money, which reduces their incentive to lend. This reduces the money supply and is typically seen as a way to slow down the economy and reduce inflation.
<human>: Write an engaging and well-written property listing description for selling a house. Address of the house and some of the details are given to you. Fill in the information gap with hallucinations if needed.
Property Address: 412 Monterey Ave, Capitola, CA 95010
Parking: 3 cars + a finished single car garage
Details:
- open floorplan
- window with views to park/mountains
- Kitchen with shaker cabinetry and white Calcatta quartz counters
<bot>: Welcome to 412 Monterey Ave, Capitola, CA 95010! This beautiful home offers an open floorplan, shaker cabinetry, and a finished single car garage. With its modern design, this house is a sight to behold. The window in the living room provides a view of the park and the mountains, bringing the outdoors in. The Kitchen features white Calcatta quartz counters and is a chef's dream. It is perfect for entertaining with its open layout and plenty of space for dining. Don't miss out on this opportunity to call this house your home!
<human>: What is artificial intelligence? Explain it so an 8 year old will understand.
<bot>: That's a great question. Artificial Intelligence (AI) is a computer program or machine that can do things that would normally require human intelligence to do. AI can do things like understand natural language, learn from experience, recognize objects, and make decisions.
Some example completions for Multilingual
Click to expand
Arabic
<human>: هل العمل نصف دوام فكرة جيدة لمراهقين بين عمر ١٥ و ١٩ عام
<bot>: نعم ، يمكن أن يكون العمل نصف دوام فكرة جيدة للمراهقين بين عمري 15 و 19 عامًا ، حيث يمكنهم من خلالها اكتساب خبرة العمل الأولى وتنمية مهارات العمل المهمة. بالإضافة إلى ذلك، يمكن أن يكون العمل نصف دوام فرصة جيدة للمراهقين لتعلم كيفية إدارة الوقت بشكل جيد.
ومع ذلك، من المهم أن تراعي الوظائف التي يتم شغلها خلال نصف الدوام احتياجات المراهقين وتجنب أي ضرر ممكن. بالإضافة إلى ذلك، من المهم أن يكون لدى المراهقين وقت فراغ كافٍ للاسترخاء والتعرف على أقرانهم وتنمية مهاراتهم الأخرى.
بالإضافة إلى ذلك، من المهم أن يدعم الآباء عمل المراهقين، ويشجعونهم على الاستمرار في الدراسة بنجاح، ويقدمون الدعم العاطفي لهم.
DISCLAIMER: Text ordering may be off for Arabic due to code block not being able to render
Chinese
<human>: 请向我推荐一部周末看的电影并说明原因。
<bot>: 我推荐的电影是疯狂动物城(Zootopia)。这部电影讲述了一个发生在动物城市里的故事。主人公是一个名叫Judy的兔子警官,她也是一个初来乍到的新人。因为初来乍到,所以她受到了其他动物同事的排挤。不过她并没有气馁,而是用自己的行动证明了自己。在解决 一个一个案子,打击罪犯的同时,她也在这个过程中收获了友情和爱情。这部电影不仅有深刻的教育意义,而且还有好看的动作和搞笑的场面。所以我觉得这是一部很适合周末看的电影。
French
<human>: Il y a t’il un petit-déjeuner de 700-1000 calories sans oeuf mais qui inclut des protéines.
<bot>:
Hindi
<human>: दिल्ली से जयपुर कैसे जाये? बस से या कार से? दोनो के बारे मे विस्तार मे बताये.
<bot>: आप दिल्ली से जयपुर कार से या बस से जा सकते है, दोनों ही तरीके आसान है।
अगर आप कार से यात्रा करते है तो यह यात्रा 4 घंटे में पूरी हो जाएगी। आपको दिल्ली से जयपुर के लिए कुछ पॉपुलर मार्ग इस तरह से है -
1. नेशनल हाई-वे 2 (NH 2) - यह दिल्ली से जयपुर जाने का सबसे पॉपुलर मार्ग है और यह पूरी तरह से एक फोर-लेन मार्ग है। इस मार्ग पर आप बिना किसी रोक-टोक के दिल्ली से जयपुर तक पहुंच सकते है।
2. दिल्ली-जयपुर मार्ग (Delhi-Jaipur Highway) - यह दिल्ली से जयपुर जाने का एक और आसान मार्ग है। यह जयपुर से दिल्ली की और जाने वाले NH 2 को जयपुर में दिल्ली-जयपुर राष्ट्रीय राजमार्ग (Delhi-Jaipur National Highway) कहते है।
3. दिल्ली-जयपुर मार्ग (Delhi-Jaipur Highway) - यह NH 2 दिल्ली से जयपुर तक जाने का एक और मार्ग है। यह दिल्ली से जयपुर की और जाने वाले NH 2 से जुड़ा हुआ है और दिल्ली से जयपुर तक सीधा जाता है।
4. दिल्ली-जयपुर मार्ग (Delhi-Jaipur Highway) - यह NH 2 दिल्ली से जयपुर तक जाने का एक और मार्ग है। यह दिल्ली से जयपुर की और जाने वाले NH 2 से जुड़ा हुआ है और दिल्ली से जयपुर तक सीधा जाता है।
जयपुर जाने के लिए आप इनमें से किसी भी मार्ग का चुनाव कर सकते है और आसानी से जयपुर पहुंच सकते है।
जयपुर जाने के लिए आप सरकारी बसों का भी चुनाव कर सकते है। दिल्ली से जयपुर के लिए कई बसों की सुविधा उपलब्ध है। दिल्ली से जयपुर के लिए आप दिल्ली के सराय रोहिल्ला बस अड्डे से आसानी से बस ले सकते है। सराय रोहिल्ला से जयपुर के लिए कई बसें उपलब्ध है जो आपको जयपुर पहुंचा देगी।
जयपुर जाने के लिए आप प्राइवेट कैब का भी उपयोग कर सकते है। यह सुविधा आप ऑनलाइन भी ले सकते है। कैब से दिल्ली से जयपुर जाने में आपको लगभग 4 घंटे लगेंगे।
Spanish
<human>: ¿Me puedes dar algunas recomendaciones de juegos? Estoy buscando juegos de rol con un sistema de magia único, un sistema de construcción único y una buena historia. Por favor limítate a juegos que se puedan jugar en PC directamente, o por emulación.
<bot>: Claro, aquí están algunas opciones que podrían interesarte:
1.- "Spellstorm": Este juego de rol tiene un sistema de magia único que se basa en la manipulación de los elementos y la creación de tormentas de magia. El sistema de construcción es basado en habilidades y la historia es interesante, con una ambientación de fantasía oscura.
2.- "Infinity Engine": Este es un sistema de juego de rol creado por "Rol Team B", que cuenta con varios títulos de calidad. El sistema de magia es único y se basa en la manipulación de portales interdimensionales. El sistema de construcción es basado en habilidades y la historia es interesante, con una ambientación de fantasía.
3.- "Arcana Evolved": Este juego de rol tiene un sistema de magia único que se basa en la manipulación de las leyes fundamentales del universo. El sistema de construcción es basado en habilidades y atributos, y la historia es interesante, con una ambientación de ciencia ficción.
4.- "Fate": Este es un sistema de juego de rol que se enfoca en la narrativa y la toma de decisiones. El sistema de magia es inexistente, pero la historia es muy interesante, con una ambientación de fantasía y ciencia ficción.
Espero que estas opciones te sean de ayuda.
Training Details
Click to expand
Training Data
Long Sequence Pretraining Data
- mc4 3.1.0
- RefinedWeb
- StarCoder
- The PILE
- Pile of Law
- 10-Ks and 10-Qs from EDGAR
- arXiv papers from Economics and Quantitative Finance
- YouTube transcripts from finance-related public YouTube videos
Finetuning Data
Training Procedure
We trained BLOOMChat-v2 with SambaNova DataScale systems with SambaNova's in-house Reconfigurable Dataflow Unit (RDU). We started from BLOOM (176B), an open-source multilingual LLM pretrained by the BigScience group. We then continued pretraining the model on an in-house mix of multilingual text (see the full breakdown above).
We instruction-tune the resulting model on OpenChatKit with each data source subsampled to 100k for one epoch, followed by sixteen epochs over the combined OpenChatKit and Dolly 2.0. All of the code used to prepare the finetuning datasets and the scripts to run finetuning and inference are open-sourced and freely available at sambanova/bloomchat
Prompting Style Used For Training
<human>: {input1 that the user wants from the bot}
<bot>: {response1}</s>
<human>: {input2 that the user wants from the bot}
<bot>: {response2}</s>
Hyperparameters
Long-sequence Pretraining
- Hardware: SambaNova Reconfigurable Dataflow Unit (RDU)
- Optimizer: AdamW
- Grad accumulation: 1
- Steps: 1620
- Global Batch size: 2048
- Batch tokens: 2048 * 8192 = 16,777,216 tokens
- Learning Rate: 6e-6
- Learning Rate Scheduler: Flat
- Warmup Steps: 0
- Weight decay: 0.1
Instruction-tuned Training on OIG
- Hardware: SambaNova Reconfigurable Dataflow Unit (RDU)
- Optimizer: AdamW
- Grad accumulation: 1
- Epochs: 1
- Global Batch size: 128
- Batch tokens: 128 * 8192 = 1,048,576 tokens
- Learning Rate: 6e-6
- Learning Rate Scheduler: Cosine Schedule with Warmup
- Warmup Steps: 0
- End Learning Ratio: 0.1
- Weight decay: 0.1
Instruction-tuned Training on Dolly 2.0 and Oasst1
- Hardware: SambaNova Reconfigurable Dataflow Unit (RDU)
- Optimizer: AdamW
- Grad accumulation: 1
- Epochs: 16
- Global Batch size: 128
- Batch tokens: 128 * 8192 = 1,048,576 tokens
- Learning Rate: 6e-6
- Learning Rate Scheduler: Cosine Schedule with Warmup
- Warmup Steps: 0
- End Learning Ratio: 0.1
- Weight decay: 0.1
Bias, Risks, and Limitations
Like all LLMs, BLOOMChat has certain limitations:
- Hallucination: BLOOMChat may sometimes generate responses that contain plausible-sounding but factually incorrect or irrelevant information.
- Code Switching: The model might unintentionally switch between languages or dialects within a single response, affecting the coherence and understandability of the output.
- Repetition: BLOOMChat may produce repetitive phrases or sentences, leading to less engaging and informative responses.
- Coding and Math: The model's performance in generating accurate code or solving complex mathematical problems may be limited.
- Toxicity: BLOOMChat may inadvertently generate responses containing inappropriate or harmful content.
Acknowledgments
We would like to extend our gratitude to Together for their contributions to BLOOMChat-v1, without which BLOOMChat-v2 would not have been possible.
We are grateful to the various researchers and open-source projects that have contributed to the development of BLOOMChat-v2. We thank BigScience for providing the BLOOM model, which served as the base for BLOOMChat-v2. For our long-sequence pretraining data: we thank Common Crawl, Google Research, and Allen Institute for AI for their contributions in making mc4-3.1.0 (avilable on HuggingFace here) possible. We thank the Technology Innovation Institute for Falcon RefinedWeb. We thank BigCode for StarCoderData, EleutherAI for The PILE, the authors of Pile of Law for their dataset, and the authors of sec-edgar for their convenient Python library for downloading SEC filings.
For our finetuning data, we thank LAION for their OIG dataset, OpenAssistant Conversations Dataset (OASST1) and also thank Databricks for providing Dolly 2.0.
We appreciate lm-eval-harness and BigScience for their essential benchmarking contributions, which is very helpful in evaluating BLOOMChat's performance. We appreciate the inspiration from the wave of various recent open-source chat models, including OpenAssistant-30B, LLaMA-Adapter-V2-65B, Vicuna-13b, Koala-13b, OASST-Pythia-12b, Alpaca-13b, ChatGLM-6b, FastChat-T5-3b, Dolly-v2-12b, LLaMA-13b, StableLM-Tuned-Alpha-7b, RedPajama-INCITE-Chat-7B-v0.1, RedPajama-INCITE-Chat-3B-v1, MPT-7B-Chat and so on. We look forward to witnessing the continued growth and success of open-source chat-based models.
We highly appreciate the hard work and dedication of these researchers and organizations towards the advancement of the open-source community. Their contributions were invaluable in the development of BLOOMChat, and we hope that our model can contribute to further advancements in the field.
Cite BLOOMChat
@software{bloomchat-v2,
title = {{BLOOMChat-v2: an Open Multilingual Chat LLM for Long Sequences}},
author = {SambaNova Systems},
url = {https://huggingface.co/sambanovasystems/BLOOMChat-176B-v2}
month = {2},
year = {2024},
version = {2.0},
}
- Downloads last month
- 18