Upload LlamaForCausalLM
Browse files- README.md +3 -121
- config.json +55 -0
- generation_config.json +9 -0
- model-00001-of-00015.safetensors +3 -0
- model-00002-of-00015.safetensors +3 -0
- model-00003-of-00015.safetensors +3 -0
- model-00004-of-00015.safetensors +3 -0
- model-00005-of-00015.safetensors +3 -0
- model-00006-of-00015.safetensors +3 -0
- model-00007-of-00015.safetensors +3 -0
- model-00008-of-00015.safetensors +3 -0
- model-00009-of-00015.safetensors +3 -0
- model-00010-of-00015.safetensors +3 -0
- model-00011-of-00015.safetensors +3 -0
- model-00012-of-00015.safetensors +3 -0
- model-00013-of-00015.safetensors +3 -0
- model-00014-of-00015.safetensors +3 -0
- model-00015-of-00015.safetensors +3 -0
- model.safetensors.index.json +0 -0
README.md
CHANGED
@@ -1,121 +1,3 @@
|
|
1 |
-
---
|
2 |
-
license: mit
|
3 |
-
|
4 |
-
- deepseek-ai/DeepSeek-R1-Distill-Llama-70B
|
5 |
-
---
|
6 |
-
|
7 |
-
BCCard/DeepSeek-R1-Distill-Llama-8B-BCGPT is a quantized model based on DeepSeek-R1-Distill-Llama-70B.
|
8 |
-
|
9 |
-
```
|
10 |
-
model.eval()
|
11 |
-
instruction = "서울의 유명한 관광 코스를 상세하게 일주일간 일별로 만들어줄래?"
|
12 |
-
|
13 |
-
messages = [
|
14 |
-
{"role": "system", "content": f"당신은 훌륭한 AI 한국 전문가입니다. 한국어로 답변해주세요."} ,
|
15 |
-
{"role": "user", "content": f"{instruction}"}
|
16 |
-
]
|
17 |
-
|
18 |
-
input_ids = tokenizer.apply_chat_template(
|
19 |
-
messages,
|
20 |
-
add_generation_prompt=True,
|
21 |
-
return_tensors="pt"
|
22 |
-
).to(model.device)
|
23 |
-
|
24 |
-
terminators = [
|
25 |
-
tokenizer.eos_token_id,
|
26 |
-
tokenizer.convert_tokens_to_ids("<|eot_id|>")
|
27 |
-
]
|
28 |
-
|
29 |
-
outputs = model.generate(
|
30 |
-
input_ids,
|
31 |
-
max_new_tokens=4096,
|
32 |
-
eos_token_id=terminators,
|
33 |
-
do_sample=True,
|
34 |
-
temperature=1.0,
|
35 |
-
top_p=0.9,
|
36 |
-
)
|
37 |
-
|
38 |
-
print(tokenizer.decode(outputs[0][input_ids.shape[-1]:], skip_special_tokens=True))
|
39 |
-
```
|
40 |
-
|
41 |
-
|
42 |
-
```
|
43 |
-
<think>
|
44 |
-
|
45 |
-
</think>
|
46 |
-
|
47 |
-
|
48 |
-
물론 서울의 유명한 관광 코스는 서울의 대표적인 관광명소들을 모아 놓은 코스로, 보통 1일, 3일~4일 정도의 일정으로 구성되어 있습니다. 아래는 대표적인 서울 관광 코스를 1일, 3일~4일 단위로 나누어 설명드리겠습니다.
|
49 |
-
|
50 |
-
### 1일 코스
|
51 |
-
**명소:** 광화문 일대, 서울의 대표적인 명소를 하루 동안 방문할 수 있는 코스입니다.
|
52 |
-
1. **경복궁**
|
53 |
-
- 서울의 대표적인 궁전으로, 전통 한옥 양식으로 유명한 곳입니다.
|
54 |
-
2. **북촌 한옥마을**
|
55 |
-
- 전통과 현대적인 예술을 함께 즐길 수 있는 공간입니다.
|
56 |
-
3. **명동 거리**
|
57 |
-
- 서울의 역사를 느끼고 쇼핑을 즐길 수 있는 활기찬 거리입니다.
|
58 |
-
|
59 |
-
### 3일~4일 코스
|
60 |
-
**명소:** 서울의 더 많은 명소를 방문하기 위한 3일~4일 일정의 코스입니다.
|
61 |
-
1. **광화문**
|
62 |
-
- 서울의 중심부로, 역사적 의미가 깊은 장소입니다.
|
63 |
-
2. **인사동**
|
64 |
-
- 전통과 현대 예술이 공존하는 거리로, 다양한 문화 체험이 가능합니다.
|
65 |
-
3. **명동 거리**
|
66 |
-
- 서울의 역사를 배우고 쇼핑을 즐길 수 있는 대표적인 장소입니다.
|
67 |
-
4. **동대문**
|
68 |
-
- 서울의 문화와 역사를 느낄 수 있는 활기찬 지역입니다.
|
69 |
-
|
70 |
-
### 상세 일정
|
71 |
-
1. **경복궁** (1일차)
|
72 |
-
- 위치: 서울 종로구
|
73 |
-
- 관람 시간: 09:00 ~ 10:00
|
74 |
-
- 특징: 경복궁의 아름다운 경관과 역사적인 배경을 감상할 수 있습니다.
|
75 |
-
|
76 |
-
2. **북촌 한옥마을** (1일차)
|
77 |
-
- 위치: 서울 종로구
|
78 |
-
- 관람 시간: 10:30 ~ 12:00
|
79 |
-
- 특징: 전통 한옥과 현대 예술이 어우러진 독특한 분위기를 경험할 수 있습니다.
|
80 |
-
|
81 |
-
3. **명동 거리** (2일차)
|
82 |
-
- 위치: 서울 중구
|
83 |
-
- 관람 시간: 13:00 ~ 15:00
|
84 |
-
- 특징: 활기찬 거리에서 다양한 쇼핑과 먹거리를 즐길 수 있습니다.
|
85 |
-
|
86 |
-
4. **동대문** (3일차)
|
87 |
-
- 위치: 서울 종로구
|
88 |
-
- 관람 시간: 15:30 ~ 17:30
|
89 |
-
- 특징: 서울의 문화와 현대적인 디자인이 조화를 이루는 공간입니다.
|
90 |
-
|
91 |
-
|
92 |
-
이처럼 서울의 유명한 관광 코스를 1일 또는 3일~4일 일정으로 나누어 방문할 수 있습니다. 각 명소의 매력을 즐기면서 서울의 다양한 문화와 역사를 경험해 보세요.
|
93 |
-
```
|
94 |
-
|
95 |
-
- Question
|
96 |
-
```
|
97 |
-
Suppose I borrowed 3 million won. A month later, I have 3 million won available to repay the loan. However, instead of paying it back, I'm considering investing it in stocks as of December 2022. What do you think about that?
|
98 |
-
```
|
99 |
-
|
100 |
-
- Answer
|
101 |
-
```
|
102 |
-
<think>
|
103 |
-
|
104 |
-
</think>
|
105 |
-
|
106 |
-
</think>
|
107 |
-
|
108 |
-
If you borrowed 3 million won and after a month you have the same amount available to repay the loan, it's important to consider the implications of not repaying the loan. Instead of repaying the loan, investing the funds in stocks carries significant risks, including market volatility and potential losses.
|
109 |
-
|
110 |
-
Here are some key points to consider:
|
111 |
-
|
112 |
-
1. **Interest and Fees**: By not repaying the loan, you may incur interest charges and potential late fees, which could increase the amount you owe over time.
|
113 |
-
|
114 |
-
2. **Risk of Loss in Stocks**: The stock market can be unpredictable, and there's a risk that the value of your investment could decrease, leading to potential losses.
|
115 |
-
|
116 |
-
3. **Legal and Credit Implications**: Failing to repay a loan can damage your credit score and may lead to legal consequences, depending on the terms of the loan and the lender's policies.
|
117 |
-
|
118 |
-
4. **Opportunity Cost**: While investing might offer the potential for gains, it's important to weigh this against the guaranteed costs of not repaying the loan.
|
119 |
-
|
120 |
-
It's generally advisable to prioritize repaying debts over speculative investments, especially if the loan terms are unfavorable or if the interest rate is high. If you're considering investing, ensure you have a clear understanding of the risks involved and consider consulting with a financial advisor to make an informed decision.
|
121 |
-
```
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
config.json
ADDED
@@ -0,0 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "DeepSeek-R1-Distill-Llama-70B",
|
3 |
+
"architectures": [
|
4 |
+
"LlamaForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_bias": false,
|
7 |
+
"attention_dropout": 0.0,
|
8 |
+
"bos_token_id": 128000,
|
9 |
+
"eos_token_id": [
|
10 |
+
128001,
|
11 |
+
128008,
|
12 |
+
128009
|
13 |
+
],
|
14 |
+
"head_dim": 128,
|
15 |
+
"hidden_act": "silu",
|
16 |
+
"hidden_size": 8192,
|
17 |
+
"initializer_range": 0.02,
|
18 |
+
"intermediate_size": 28672,
|
19 |
+
"max_position_embeddings": 131072,
|
20 |
+
"mlp_bias": false,
|
21 |
+
"model_type": "llama",
|
22 |
+
"num_attention_heads": 64,
|
23 |
+
"num_hidden_layers": 80,
|
24 |
+
"num_key_value_heads": 8,
|
25 |
+
"pretraining_tp": 1,
|
26 |
+
"quantization_config": {
|
27 |
+
"_load_in_4bit": false,
|
28 |
+
"_load_in_8bit": true,
|
29 |
+
"bnb_4bit_compute_dtype": "float32",
|
30 |
+
"bnb_4bit_quant_storage": "uint8",
|
31 |
+
"bnb_4bit_quant_type": "fp4",
|
32 |
+
"bnb_4bit_use_double_quant": false,
|
33 |
+
"llm_int8_enable_fp32_cpu_offload": false,
|
34 |
+
"llm_int8_has_fp16_weight": false,
|
35 |
+
"llm_int8_skip_modules": null,
|
36 |
+
"llm_int8_threshold": 6.0,
|
37 |
+
"load_in_4bit": false,
|
38 |
+
"load_in_8bit": true,
|
39 |
+
"quant_method": "bitsandbytes"
|
40 |
+
},
|
41 |
+
"rms_norm_eps": 1e-05,
|
42 |
+
"rope_scaling": {
|
43 |
+
"factor": 8.0,
|
44 |
+
"high_freq_factor": 4.0,
|
45 |
+
"low_freq_factor": 1.0,
|
46 |
+
"original_max_position_embeddings": 8192,
|
47 |
+
"rope_type": "llama3"
|
48 |
+
},
|
49 |
+
"rope_theta": 500000.0,
|
50 |
+
"tie_word_embeddings": false,
|
51 |
+
"torch_dtype": "bfloat16",
|
52 |
+
"transformers_version": "4.48.2",
|
53 |
+
"use_cache": true,
|
54 |
+
"vocab_size": 128256
|
55 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 128000,
|
4 |
+
"do_sample": true,
|
5 |
+
"eos_token_id": 128001,
|
6 |
+
"temperature": 0.6,
|
7 |
+
"top_p": 0.95,
|
8 |
+
"transformers_version": "4.48.2"
|
9 |
+
}
|
model-00001-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:81a2023952146d1ece44cdc893616cf0c2fa3efd1557cdea67bc89f9b40e6e50
|
3 |
+
size 4820443929
|
model-00002-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4433056377edcb0ed67701128346d0da64d5f7925f0065065c63a12f982c3636
|
3 |
+
size 4984984974
|
model-00003-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b825d7109c787df2b2811c5741b3b07bf36a8ac164b1ca3b8bf1c1122cbe456e
|
3 |
+
size 4901107969
|
model-00004-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8a6dd53c70d35b0a6de7dc0f7f9cdc962377d42c70d1e8066b86ce571dda51b0
|
3 |
+
size 4901059073
|
model-00005-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4946bc0223e56f8cb36eda54955da58318c05823c43e95f4f8f9833cd0d65fde
|
3 |
+
size 4901059073
|
model-00006-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b9248732a6e7c800fcad9f83acd077354a2252e6914ea7eaac1059c6c75b4530
|
3 |
+
size 4984985102
|
model-00007-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7bd3a5fbc22562b899dc0210a266a4db6024a443001050c383f7298273673409
|
3 |
+
size 4901107993
|
model-00008-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee0afec835fcd6e36247bebaa9c05b7575b1de689c377b85231795991bb01ca4
|
3 |
+
size 4901059073
|
model-00009-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a03b9978e7cfb1697d7717de532b5afac941605f3c2215b03c27a9f4e9592d42
|
3 |
+
size 4901059073
|
model-00010-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:591673fbaa6a6eb8de835bff4d421a25a4a1bc9525cf7f14cd393e0a8f8c22ad
|
3 |
+
size 4984985102
|
model-00011-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b07957184cfcaee5f8898893c10ecd5bf7ee4bbc52ee1e0695c86f650e2e5cce
|
3 |
+
size 4901107993
|
model-00012-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4d9db6a46be2d2bfa6617b8f82b21506352d6db0bad44f3d78453a9c451f5659
|
3 |
+
size 4901059073
|
model-00013-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3320c119c8ea2f586078dac894c778c992ce33d723ea692b59098f09902f7f95
|
3 |
+
size 4901059073
|
model-00014-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:590751c92bdd7e09ebbea8c52ccd4b6e1199032f91606f9fb7ffdbf4bd80bb93
|
3 |
+
size 4984985102
|
model-00015-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:505d8e58668ad57fcbf054b276acdc5fe04809cbcab5794dcb4688735b16c4c1
|
3 |
+
size 3813381278
|
model.safetensors.index.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|