File size: 727 Bytes
ccdd9eb
 
 
 
 
 
abe2ed3
8ed3395
4f0d53e
 
 
 
147b85c
ccdd9eb
63b69bf
 
 
ccdd9eb
 
 
 
 
dccfb50
ccdd9eb
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
---
license: apache-2.0
language:
- en
---

*[December 2023]* We are ranked *4th* on the overall leaderboard and **1st** in the 7B leaderboard! 
🔥🔥🔥🔥

![image/png](https://cdn-uploads.huggingface.co/production/uploads/63c06fba8d1175e3399c16e6/yEPpr0V-D9V4m1a2pMuQs.png)

DPO from the model ranked *6th* on the overall leaderboard and **1st** in the 7B leaderboard (12th December 2023) - v1olet/v1olet_marcoroni-go-bruins-merge-7B 
https://huggingface.co/v1olet/v1olet_marcoroni-go-bruins-merge-7B

Training data:
comparison_gpt4_en,en_orca_dpo

Train for 1 epoch.

You can use alpaca template.
```
template_format = """{system}

### Instruction:
{prompt}

### Response:
"""
```

Developed by: Trong-Hieu Nguyen-Mau