update model card, nice :)
Browse files
README.md
CHANGED
@@ -64,7 +64,8 @@ model-index:
|
|
64 |
value: 64.68
|
65 |
name: normalized accuracy
|
66 |
source:
|
67 |
-
url:
|
|
|
68 |
name: Open LLM Leaderboard
|
69 |
- task:
|
70 |
type: text-generation
|
@@ -80,7 +81,8 @@ model-index:
|
|
80 |
value: 83.75
|
81 |
name: normalized accuracy
|
82 |
source:
|
83 |
-
url:
|
|
|
84 |
name: Open LLM Leaderboard
|
85 |
- task:
|
86 |
type: text-generation
|
@@ -97,7 +99,8 @@ model-index:
|
|
97 |
value: 62.31
|
98 |
name: accuracy
|
99 |
source:
|
100 |
-
url:
|
|
|
101 |
name: Open LLM Leaderboard
|
102 |
- task:
|
103 |
type: text-generation
|
@@ -113,7 +116,8 @@ model-index:
|
|
113 |
- type: mc2
|
114 |
value: 55.15
|
115 |
source:
|
116 |
-
url:
|
|
|
117 |
name: Open LLM Leaderboard
|
118 |
- task:
|
119 |
type: text-generation
|
@@ -130,7 +134,8 @@ model-index:
|
|
130 |
value: 76.24
|
131 |
name: accuracy
|
132 |
source:
|
133 |
-
url:
|
|
|
134 |
name: Open LLM Leaderboard
|
135 |
- task:
|
136 |
type: text-generation
|
@@ -147,8 +152,11 @@ model-index:
|
|
147 |
value: 57.62
|
148 |
name: accuracy
|
149 |
source:
|
150 |
-
url:
|
|
|
151 |
name: Open LLM Leaderboard
|
|
|
|
|
152 |
---
|
153 |
|
154 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/U0zyXVGj-O8a7KP3BvPue.png)
|
@@ -311,17 +319,6 @@ Quantizationed versions of this model is available.
|
|
311 |
|
312 |
- https://huggingface.co/bartowski/Einstein-v4-7B-exl2
|
313 |
|
314 |
-
You can switch up branches in the repo to use the one you want
|
315 |
-
|
316 |
-
| Branch | Bits | lm_head bits | VRAM (4k) | VRAM (16k) | VRAM (32k) | Description |
|
317 |
-
| ----- | ---- | ------- | ------ | ------ | ------ | ------------ |
|
318 |
-
| [8_0](https://huggingface.co/bartowski/Einstein-v4-7B-exl2/tree/8_0) | 8.0 | 8.0 | 8.4 GB | 9.8 GB | 11.8 GB | Maximum quality that ExLlamaV2 can produce, near unquantized performance. |
|
319 |
-
| [6_5](https://huggingface.co/bartowski/Einstein-v4-7B-exl2/tree/6_5) | 6.5 | 8.0 | 7.2 GB | 8.6 GB | 10.6 GB | Very similar to 8.0, good tradeoff of size vs performance, **recommended**. |
|
320 |
-
| [5_0](https://huggingface.co/bartowski/Einstein-v4-7B-exl2/tree/5_0) | 5.0 | 6.0 | 6.0 GB | 7.4 GB | 9.4 GB | Slightly lower quality vs 6.5, but usable on 8GB cards. |
|
321 |
-
| [4_25](https://huggingface.co/bartowski/Einstein-v4-7B-exl2/tree/4_25) | 4.25 | 6.0 | 5.3 GB | 6.7 GB | 8.7 GB | GPTQ equivalent bits per weight, slightly higher quality. |
|
322 |
-
| [3_5](https://huggingface.co/bartowski/Einstein-v4-7B-exl2/tree/3_5) | 3.5 | 6.0 | 4.7 GB | 6.1 GB | 8.1 GB | Lower quality, only use if you have to. |
|
323 |
-
|
324 |
-
|
325 |
# 🎯 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
326 |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v4-7B)
|
327 |
|
@@ -335,6 +332,22 @@ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-le
|
|
335 |
|Winogrande (5-shot) |76.24|
|
336 |
|GSM8k (5-shot) |57.62|
|
337 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
338 |
# 🤖 Additional information about training
|
339 |
|
340 |
This model is full fine-tuned for 1.5 epoch.
|
@@ -360,4 +373,4 @@ Thanks to all open source AI community.
|
|
360 |
|
361 |
If you would like to support me:
|
362 |
|
363 |
-
[☕ Buy Me a Coffee](https://www.buymeacoffee.com/weyaxi)
|
|
|
64 |
value: 64.68
|
65 |
name: normalized accuracy
|
66 |
source:
|
67 |
+
url: >-
|
68 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
|
69 |
name: Open LLM Leaderboard
|
70 |
- task:
|
71 |
type: text-generation
|
|
|
81 |
value: 83.75
|
82 |
name: normalized accuracy
|
83 |
source:
|
84 |
+
url: >-
|
85 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
|
86 |
name: Open LLM Leaderboard
|
87 |
- task:
|
88 |
type: text-generation
|
|
|
99 |
value: 62.31
|
100 |
name: accuracy
|
101 |
source:
|
102 |
+
url: >-
|
103 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
|
104 |
name: Open LLM Leaderboard
|
105 |
- task:
|
106 |
type: text-generation
|
|
|
116 |
- type: mc2
|
117 |
value: 55.15
|
118 |
source:
|
119 |
+
url: >-
|
120 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
|
121 |
name: Open LLM Leaderboard
|
122 |
- task:
|
123 |
type: text-generation
|
|
|
134 |
value: 76.24
|
135 |
name: accuracy
|
136 |
source:
|
137 |
+
url: >-
|
138 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
|
139 |
name: Open LLM Leaderboard
|
140 |
- task:
|
141 |
type: text-generation
|
|
|
152 |
value: 57.62
|
153 |
name: accuracy
|
154 |
source:
|
155 |
+
url: >-
|
156 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
|
157 |
name: Open LLM Leaderboard
|
158 |
+
language:
|
159 |
+
- en
|
160 |
---
|
161 |
|
162 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/U0zyXVGj-O8a7KP3BvPue.png)
|
|
|
319 |
|
320 |
- https://huggingface.co/bartowski/Einstein-v4-7B-exl2
|
321 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
322 |
# 🎯 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
323 |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v4-7B)
|
324 |
|
|
|
332 |
|Winogrande (5-shot) |76.24|
|
333 |
|GSM8k (5-shot) |57.62|
|
334 |
|
335 |
+
# 📚 Some resources, discussions and reviews aboout this model
|
336 |
+
|
337 |
+
#### 🐦 Announcement tweet:
|
338 |
+
|
339 |
+
https://twitter.com/Weyaxi/status/1765851433448944125
|
340 |
+
|
341 |
+
#### 🔍 Reddit post in r/LocalLLaMA:
|
342 |
+
|
343 |
+
- https://www.reddit.com/r/LocalLLaMA/comments/1b9gmvl/meet_einsteinv47b_mistralbased_sft_model_using/
|
344 |
+
|
345 |
+
#### ▶️ Youtube Videos
|
346 |
+
|
347 |
+
- https://www.youtube.com/watch?v=-3YWgHJIORE&t=18s
|
348 |
+
|
349 |
+
- https://www.youtube.com/watch?v=Xo2ySU8gja0
|
350 |
+
|
351 |
# 🤖 Additional information about training
|
352 |
|
353 |
This model is full fine-tuned for 1.5 epoch.
|
|
|
373 |
|
374 |
If you would like to support me:
|
375 |
|
376 |
+
[☕ Buy Me a Coffee](https://www.buymeacoffee.com/weyaxi)
|