Commit
•
34197ae
1
Parent(s):
e2232ed
Update README.md
Browse files
README.md
CHANGED
@@ -25,8 +25,8 @@ license: apache-2.0
|
|
25 |
Notus is going to be a collection of fine-tuned models using DPO, similarly to Zephyr, but mainly focused
|
26 |
on the Direct Preference Optimization (DPO) step, aiming to incorporate preference feedback into the LLMs
|
27 |
when fine-tuning those. Notus models are intended to be used as assistants via chat-like applications, and
|
28 |
-
are evaluated with the MT-Bench and
|
29 |
-
also using DPO.
|
30 |
|
31 |
## Model Details
|
32 |
|
|
|
25 |
Notus is going to be a collection of fine-tuned models using DPO, similarly to Zephyr, but mainly focused
|
26 |
on the Direct Preference Optimization (DPO) step, aiming to incorporate preference feedback into the LLMs
|
27 |
when fine-tuning those. Notus models are intended to be used as assistants via chat-like applications, and
|
28 |
+
are evaluated with the MT-Bench, AlpacaEval, and LM Evaluation Harness benchmarks, to be directly compared
|
29 |
+
with Zephyr fine-tuned models also using DPO.
|
30 |
|
31 |
## Model Details
|
32 |
|