Update README.md
Browse files
README.md
CHANGED
@@ -18,7 +18,7 @@ pipeline_tag: text-generation
|
|
18 |
|
19 |
Welcome to **GenZ**, an advanced Large Language Model (LLM) fine-tuned on the foundation of Meta's open-source Llama V2 70B parameter model. At Bud Ecosystem, we believe in the power of open-source collaboration to drive the advancement of technology at an accelerated pace. Our vision is to democratize access to fine-tuned LLMs, and to that end, we will be releasing a series of models across different parameter counts (7B, 13B, and 70B) and quantizations (32-bit and 4-bit) for the open-source community to use, enhance, and build upon.
|
20 |
|
21 |
-
<p align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/GenZ/main/assets/
|
22 |
|
23 |
The smaller quantization version of our models makes them more accessible, enabling their use even on personal computers. This opens up a world of possibilities for developers, researchers, and enthusiasts to experiment with these models and contribute to the collective advancement of language model technology.
|
24 |
|
@@ -47,6 +47,28 @@ And this isn't the end. It's just the beginning of a journey towards creating mo
|
|
47 |
|
48 |
---
|
49 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
50 |
|
51 |
<h2>Getting Started on Hugging Face π€</h2>
|
52 |
|
@@ -157,29 +179,6 @@ Here are the hyperparameters we used for fine-tuning:
|
|
157 |
| Precision | FP16 |
|
158 |
| Optimizer | AdamW |
|
159 |
|
160 |
-
---
|
161 |
-
|
162 |
-
<h2>Evaluations π―</h2>
|
163 |
-
|
164 |
-
Evaluating our model is a key part of our fine-tuning process. It helps us understand how our model is performing and how it stacks up against other models. Here's a look at some of the key evaluations for GenZ 70B:
|
165 |
-
|
166 |
-
<h3>Benchmark Comparison</h3>
|
167 |
-
|
168 |
-
We've compared GenZ models to understand the improvements our fine-tuning has achieved.
|
169 |
-
|
170 |
-
| Model Name | MT Bench | MMLU | Human Eval | Hellaswag | BBH |
|
171 |
-
|:----------:|:--------:|:----:|:----------:|:---------:|:----:|
|
172 |
-
| Genz 13B | 6.12 | 53.62| 17.68 | 77.38 | 37.76|
|
173 |
-
| Genz 13B v2| 6.79 | 53.68| 21.95 | 77.48 | 38.1 |
|
174 |
-
| Genz 70B | 7.33 | 70.32| | | |
|
175 |
-
|
176 |
-
<h3>MT Bench Score</h3>
|
177 |
-
|
178 |
-
A key evaluation metric we use is the MT Bench score. This score provides a comprehensive assessment of our model's performance across a range of tasks.
|
179 |
-
|
180 |
-
<p align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/GenZ/main/assets/mt_bench_score.png" width="500"></p>
|
181 |
-
|
182 |
-
|
183 |
---
|
184 |
|
185 |
<h2>Looking Ahead π</h2>
|
|
|
18 |
|
19 |
Welcome to **GenZ**, an advanced Large Language Model (LLM) fine-tuned on the foundation of Meta's open-source Llama V2 70B parameter model. At Bud Ecosystem, we believe in the power of open-source collaboration to drive the advancement of technology at an accelerated pace. Our vision is to democratize access to fine-tuned LLMs, and to that end, we will be releasing a series of models across different parameter counts (7B, 13B, and 70B) and quantizations (32-bit and 4-bit) for the open-source community to use, enhance, and build upon.
|
20 |
|
21 |
+
<p align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/GenZ/main/assets/mt_bench_compare" width="500"></p>
|
22 |
|
23 |
The smaller quantization version of our models makes them more accessible, enabling their use even on personal computers. This opens up a world of possibilities for developers, researchers, and enthusiasts to experiment with these models and contribute to the collective advancement of language model technology.
|
24 |
|
|
|
47 |
|
48 |
---
|
49 |
|
50 |
+
<h2>Evaluations π―</h2>
|
51 |
+
|
52 |
+
Evaluating our model is a key part of our fine-tuning process. It helps us understand how our model is performing and how it stacks up against other models. Here's a look at some of the key evaluations for GenZ 70B:
|
53 |
+
|
54 |
+
<h3>Benchmark Comparison</h3>
|
55 |
+
|
56 |
+
We've compared GenZ models to understand the improvements our fine-tuning has achieved.
|
57 |
+
|
58 |
+
| Model Name | MT Bench | MMLU | Human Eval | BBH |
|
59 |
+
|:----------:|:--------:|:----:|:----------:|:----:|
|
60 |
+
| Genz 13B | 6.12 | 53.62| 17.68 | 37.76|
|
61 |
+
| Genz 13B v2| 6.79 | 53.68| 21.95 | 38.1 |
|
62 |
+
| Genz 70B | 7.33 | 70.32| 37.8 |54.69 |
|
63 |
+
|
64 |
+
<h3>MT Bench Score</h3>
|
65 |
+
|
66 |
+
A key evaluation metric we use is the MT Bench score. This score provides a comprehensive assessment of our model's performance across a range of tasks.
|
67 |
+
|
68 |
+
<p align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/GenZ/main/assets/mt_bench_score.png" width="500"></p>
|
69 |
+
|
70 |
+
|
71 |
+
---
|
72 |
|
73 |
<h2>Getting Started on Hugging Face π€</h2>
|
74 |
|
|
|
179 |
| Precision | FP16 |
|
180 |
| Optimizer | AdamW |
|
181 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
182 |
---
|
183 |
|
184 |
<h2>Looking Ahead π</h2>
|