|
--- |
|
language: |
|
- en |
|
pipeline_tag: text-generation |
|
datasets: |
|
- nlpai-lab/databricks-dolly-15k-ko |
|
- kyujinpy/KOR-OpenOrca-Platypus-v3 |
|
- KETI-AIR/kor_boolq |
|
- heegyu/open-korean-instructions |
|
license: cc-by-nc-sa-4.0 |
|
--- |
|
|
|
**Input** Models input text only. |
|
|
|
**Output** Models generate text only. |
|
|
|
**Base Model** [yanolja/KoSOLAR-10.7B-v0.1](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.1-deprecated) |
|
|
|
**Training Dataset** |
|
- [nlpai-lab/databricks-dolly-15k-ko](https://huggingface.co/datasets/nlpai-lab/databricks-dolly-15k-ko) |
|
- [kyujinpy/KOR-OpenOrca-Platypus-v3](https://huggingface.co/datasets/kyujinpy/KOR-OpenOrca-Platypus-v3) |
|
- [heegyu/open-korean-instructions](heegyu/open-korean-instructions) |
|
- [KETI-AIR/kor_boolq](https://huggingface.co/datasets/KETI-AIR/kor_boolq) |
|
- [AIhub μν λ²μ λ°μ΄ν° μΌλΆ](https://aihub.or.kr/aihubdata/data/view.do?currMenu=115&topMenu=100&dataSetSn=71593) |
|
|
|
# Implementation Code |
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
import torch |
|
repo = "ifuseok/sft-solar-10.7b-v2.1-dpo" |
|
OpenOrca = AutoModelForCausalLM.from_pretrained( |
|
repo, |
|
return_dict=True, |
|
torch_dtype=torch.float16, |
|
device_map='auto' |
|
) |
|
OpenOrca_tokenizer = AutoTokenizer.from_pretrained(repo) |
|
``` |
|
|
|
# Prompt Example |
|
``` |
|
### System: |
|
μμ€ν
λ©μμ§ μ
λλ€. |
|
### User: |
|
μ μ μ
λλ€. |
|
### Assistant |
|
μ΄μμ€ν΄νΈ μ
λλ€. |
|
``` |