Update README.md
Browse files
README.md
CHANGED
@@ -11,8 +11,19 @@ tags:
|
|
11 |
# Lumina-3
|
12 |
|
13 |
Lumina-3 is a Mixture of Experts (MoE) using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing).
|
14 |
-
From my testing, it is the best model I have made so far and does well overall for its size; from roleplaying, math, and coding.
|
15 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
|
17 |
## 💻 Usage
|
18 |
|
|
|
11 |
# Lumina-3
|
12 |
|
13 |
Lumina-3 is a Mixture of Experts (MoE) using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing).
|
14 |
+
From my testing, it is the best model I have made so far and does well overall for its size; from roleplaying, math, and coding. This model uses a context window of up to 32k.
|
15 |
+
|
16 |
+
## 🏆 Open LLM Leaderboard Evaluation Results
|
17 |
+
|
18 |
+
| Metric |Value|
|
19 |
+
|---------------------------------|----:|
|
20 |
+
|Avg. |74.53|
|
21 |
+
|AI2 Reasoning Challenge (25-Shot)|71.16|
|
22 |
+
|HellaSwag (10-Shot) |87.20|
|
23 |
+
|MMLU (5-Shot) |65.52|
|
24 |
+
|TruthfulQA (0-shot) |68.25|
|
25 |
+
|Winogrande (5-shot) |82.08|
|
26 |
+
|GSM8k (5-shot) |72.93|
|
27 |
|
28 |
## 💻 Usage
|
29 |
|