nbeerbower commited on
Commit
b9ff285
·
1 Parent(s): 3d67735

model file

Browse files
Files changed (5) hide show
  1. .gitattributes +1 -0
  2. README.md +46 -1
  3. config.json +28 -0
  4. ggml-model-q4_0.gguf +3 -0
  5. mergekit_config.yml +15 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.gguf filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,48 @@
1
  ---
2
- license: apache-2.0
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model:
3
+ - BarryFutureman/WildMarcoroni-Variant1-7B
4
+ - nbeerbower/bruphin-delta
5
+ tags:
6
+ - mergekit
7
+ - merge
8
+
9
  ---
10
+ # bruphin-epsilon-GGUF-q4_0
11
+
12
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
13
+
14
+ Quantized using [llama.cpp](https://github.com/ggerganov/llama.cpp).
15
+
16
+ ## Merge Details
17
+ ### Merge Method
18
+
19
+ This model was merged using the SLERP merge method.
20
+
21
+ ### Models Merged
22
+
23
+ The following models were included in the merge:
24
+ * [BarryFutureman/WildMarcoroni-Variant1-7B](https://huggingface.co/BarryFutureman/WildMarcoroni-Variant1-7B)
25
+ * [nbeerbower/bruphin-delta](https://huggingface.co/nbeerbower/bruphin-delta)
26
+
27
+ ### Configuration
28
+
29
+ The following YAML configuration was used to produce this model:
30
+
31
+ ```yaml
32
+ models:
33
+ - model: nbeerbower/bruphin-delta
34
+ layer_range: [0, 32]
35
+ - model: BarryFutureman/WildMarcoroni-Variant1-7B
36
+ layer_range: [0, 32]
37
+ merge_method: slerp
38
+ base_model: BarryFutureman/WildMarcoroni-Variant1-7B
39
+ parameters:
40
+ t:
41
+ - filter: self_attn
42
+ value: [0, 0.5, 0.3, 0.7, 1]
43
+ - filter: mlp
44
+ value: [1, 0.5, 0.7, 0.3, 0]
45
+ - value: 0.5
46
+ dtype: bfloat16
47
+
48
+ ```
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "nbeerbower/bruphin-epsilon",
3
+ "architectures": [
4
+ "MistralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 4096,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 14336,
13
+ "max_position_embeddings": 32768,
14
+ "model_type": "mistral",
15
+ "num_attention_heads": 32,
16
+ "num_hidden_layers": 32,
17
+ "num_key_value_heads": 8,
18
+ "pad_token_id": 2,
19
+ "rms_norm_eps": 1e-05,
20
+ "rope_theta": 10000.0,
21
+ "sliding_window": 4096,
22
+ "tie_word_embeddings": false,
23
+ "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.38.0.dev0",
25
+ "unsloth_version": "2024.1",
26
+ "use_cache": true,
27
+ "vocab_size": 32000
28
+ }
ggml-model-q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7672ecdbd93c08ef0328cb75fed0fb36ace6f539b447e6bc5b8d4ab10d24546a
3
+ size 4108916928
mergekit_config.yml ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ models:
2
+ - model: nbeerbower/bruphin-delta
3
+ layer_range: [0, 32]
4
+ - model: BarryFutureman/WildMarcoroni-Variant1-7B
5
+ layer_range: [0, 32]
6
+ merge_method: slerp
7
+ base_model: BarryFutureman/WildMarcoroni-Variant1-7B
8
+ parameters:
9
+ t:
10
+ - filter: self_attn
11
+ value: [0, 0.5, 0.3, 0.7, 1]
12
+ - filter: mlp
13
+ value: [1, 0.5, 0.7, 0.3, 0]
14
+ - value: 0.5
15
+ dtype: bfloat16