Kevinkre commited on
Commit
c32af78
·
verified ·
1 Parent(s): 1178de9

(Trained with Unsloth)

Browse files
Files changed (1) hide show
  1. config.json +1 -19
config.json CHANGED
@@ -1,21 +1,3 @@
1
  {
2
- "architectures": [
3
- "PhiForCausalLM"
4
- ],
5
- "model_type": "phi",
6
- "torch_dtype": "float16",
7
- "transformers_version": "4.35.0",
8
- "max_position_embeddings": 2048,
9
- "quantization_config": {
10
- "load_in_4bit": true,
11
- "bnb_4bit_compute_dtype": "float16"
12
- },
13
- "peft_config": {
14
- "peft_type": "LORA",
15
- "task_type": "CAUSAL_LM",
16
- "inference_mode": false,
17
- "r": 16,
18
- "lora_alpha": 16,
19
- "lora_dropout": 0
20
- }
21
  }
 
1
  {
2
+ "model_type": "llama"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  }