dittops commited on
Commit
19a49d8
β€’
1 Parent(s): 05f978e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -24
README.md CHANGED
@@ -18,7 +18,7 @@ pipeline_tag: text-generation
18
 
19
  Welcome to **GenZ**, an advanced Large Language Model (LLM) fine-tuned on the foundation of Meta's open-source Llama V2 70B parameter model. At Bud Ecosystem, we believe in the power of open-source collaboration to drive the advancement of technology at an accelerated pace. Our vision is to democratize access to fine-tuned LLMs, and to that end, we will be releasing a series of models across different parameter counts (7B, 13B, and 70B) and quantizations (32-bit and 4-bit) for the open-source community to use, enhance, and build upon.
20
 
21
- <p align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/GenZ/main/assets/MTBench_CompareChart_28July2023.png" width="500"></p>
22
 
23
  The smaller quantization version of our models makes them more accessible, enabling their use even on personal computers. This opens up a world of possibilities for developers, researchers, and enthusiasts to experiment with these models and contribute to the collective advancement of language model technology.
24
 
@@ -47,6 +47,28 @@ And this isn't the end. It's just the beginning of a journey towards creating mo
47
 
48
  ---
49
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
 
51
  <h2>Getting Started on Hugging Face πŸ€—</h2>
52
 
@@ -157,29 +179,6 @@ Here are the hyperparameters we used for fine-tuning:
157
  | Precision | FP16 |
158
  | Optimizer | AdamW |
159
 
160
- ---
161
-
162
- <h2>Evaluations 🎯</h2>
163
-
164
- Evaluating our model is a key part of our fine-tuning process. It helps us understand how our model is performing and how it stacks up against other models. Here's a look at some of the key evaluations for GenZ 70B:
165
-
166
- <h3>Benchmark Comparison</h3>
167
-
168
- We've compared GenZ models to understand the improvements our fine-tuning has achieved.
169
-
170
- | Model Name | MT Bench | MMLU | Human Eval | Hellaswag | BBH |
171
- |:----------:|:--------:|:----:|:----------:|:---------:|:----:|
172
- | Genz 13B | 6.12 | 53.62| 17.68 | 77.38 | 37.76|
173
- | Genz 13B v2| 6.79 | 53.68| 21.95 | 77.48 | 38.1 |
174
- | Genz 70B | 7.33 | 70.32| | | |
175
-
176
- <h3>MT Bench Score</h3>
177
-
178
- A key evaluation metric we use is the MT Bench score. This score provides a comprehensive assessment of our model's performance across a range of tasks.
179
-
180
- <p align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/GenZ/main/assets/mt_bench_score.png" width="500"></p>
181
-
182
-
183
  ---
184
 
185
  <h2>Looking Ahead πŸ‘€</h2>
 
18
 
19
  Welcome to **GenZ**, an advanced Large Language Model (LLM) fine-tuned on the foundation of Meta's open-source Llama V2 70B parameter model. At Bud Ecosystem, we believe in the power of open-source collaboration to drive the advancement of technology at an accelerated pace. Our vision is to democratize access to fine-tuned LLMs, and to that end, we will be releasing a series of models across different parameter counts (7B, 13B, and 70B) and quantizations (32-bit and 4-bit) for the open-source community to use, enhance, and build upon.
20
 
21
+ <p align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/GenZ/main/assets/mt_bench_compare" width="500"></p>
22
 
23
  The smaller quantization version of our models makes them more accessible, enabling their use even on personal computers. This opens up a world of possibilities for developers, researchers, and enthusiasts to experiment with these models and contribute to the collective advancement of language model technology.
24
 
 
47
 
48
  ---
49
 
50
+ <h2>Evaluations 🎯</h2>
51
+
52
+ Evaluating our model is a key part of our fine-tuning process. It helps us understand how our model is performing and how it stacks up against other models. Here's a look at some of the key evaluations for GenZ 70B:
53
+
54
+ <h3>Benchmark Comparison</h3>
55
+
56
+ We've compared GenZ models to understand the improvements our fine-tuning has achieved.
57
+
58
+ | Model Name | MT Bench | MMLU | Human Eval | BBH |
59
+ |:----------:|:--------:|:----:|:----------:|:----:|
60
+ | Genz 13B | 6.12 | 53.62| 17.68 | 37.76|
61
+ | Genz 13B v2| 6.79 | 53.68| 21.95 | 38.1 |
62
+ | Genz 70B | 7.33 | 70.32| 37.8 |54.69 |
63
+
64
+ <h3>MT Bench Score</h3>
65
+
66
+ A key evaluation metric we use is the MT Bench score. This score provides a comprehensive assessment of our model's performance across a range of tasks.
67
+
68
+ <p align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/GenZ/main/assets/mt_bench_score.png" width="500"></p>
69
+
70
+
71
+ ---
72
 
73
  <h2>Getting Started on Hugging Face πŸ€—</h2>
74
 
 
179
  | Precision | FP16 |
180
  | Optimizer | AdamW |
181
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
182
  ---
183
 
184
  <h2>Looking Ahead πŸ‘€</h2>