grimjim commited on
Commit
8152881
1 Parent(s): b6db2df

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +73 -81
README.md CHANGED
@@ -1,81 +1,73 @@
1
- ---
2
- base_model:
3
- - openlynn/Llama-3-Soliloquy-8B-v2
4
- - grimjim/llama-3-Nephilim-v1-8B
5
- - grimjim/Llama-3-Instruct-8B-SPPO-Iter3-SimPO-merge
6
- - grimjim/llama-3-aaditya-OpenBioLLM-8B
7
- - tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1
8
- - MLP-KTLim/llama-3-Korean-Bllossom-8B
9
- library_name: transformers
10
- tags:
11
- - mergekit
12
- - merge
13
- license: cc-by-nc-4.0
14
- pipeline_tag: text-generation
15
- ---
16
- # grimjim/llama-3-Nephilim-v2.1-8B-GGUF
17
-
18
- This repo contains select GGUF quants of a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
19
-
20
- Will bf16 weights are [available here](https://huggingface.co/grimjim/llama-3-Nephilim-v2.1-8B).
21
-
22
- Unorthodox contributions to this merge include a biomedical model, a roleplay-focused model, a Japanese language model, and a Korean language model; the original base model was an abliterated model combined with a cybersecurity model. A merger of two recent preference optimization models was also included to shore up reasoning, with the expectation that overall good reasoning and coherence will contribute to narrative flow. Roleplay text generation appears to be easily steered toward emotionally intelligent characters. It should be noted that the vast majority of merge contributions (as measured by weight) were not intended for roleplay.
23
-
24
- Tested with 8k context length, temperature=1, and minP=0.01 using Instruct prompts for Llama 3. Example Llama 3 presets can be [downloaded here](https://huggingface.co/debased-ai/SillyTavern-settings/tree/main/presets). Specifically, "Llama 3 Lukewarm Presets 8k context.json" corresponds to test settings.
25
-
26
- Example context templates variants tested with Llama 3 can be [downloaded here](https://huggingface.co/debased-ai/SillyTavern-settings/tree/main/advanced_formatting/context_template); their corresponding Instruct prompts can be [downloaded here](https://huggingface.co/debased-ai/SillyTavern-settings/tree/main/advanced_formatting/instruct_mode). Specifically, the two "Llama 3 Instruct Safety.json" files correspond to test settings.
27
-
28
- Care should be taken when using this model, as it is possible that harmful outputs may be generated. Given that this model is derivative, responsible use is further mandated by the [WhiteRabbitNeo Usage Restrictions Extension](https://huggingface.co/WhiteRabbitNeo/Llama-3-WhiteRabbitNeo-8B-v2.0) to the Llama-3 License. This model is further subject to CC-BY-NC-4.0 by default, meaning that commercial use is restricted, barring an alternative licensing agreement.
29
-
30
- Built with Meta Llama 3.
31
-
32
- ## Merge Details
33
- ### Merge Method
34
-
35
- This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [grimjim/llama-3-Nephilim-v1-8B](https://huggingface.co/grimjim/llama-3-Nephilim-v1-8B) as a base.
36
-
37
- ### Models Merged
38
-
39
- The following models were included in the merge:
40
- * [openlynn/Llama-3-Soliloquy-8B-v2](https://huggingface.co/openlynn/Llama-3-Soliloquy-8B-v2)
41
- * [grimjim/Llama-3-Instruct-8B-SPPO-Iter3-SimPO-merge](https://huggingface.co/grimjim/Llama-3-Instruct-8B-SPPO-Iter3-SimPO-merge)
42
- * [grimjim/llama-3-aaditya-OpenBioLLM-8B](https://huggingface.co/grimjim/llama-3-aaditya-OpenBioLLM-8B)
43
- * [tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1](https://huggingface.co/tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1)
44
- * [MLP-KTLim/llama-3-Korean-Bllossom-8B](https://huggingface.co/MLP-KTLim/llama-3-Korean-Bllossom-8B)
45
-
46
- ### Configuration
47
-
48
- The following YAML configuration was used to produce this model:
49
-
50
- ```yaml
51
- base_model: grimjim/llama-3-Nephilim-v1-8B
52
- dtype: bfloat16
53
- merge_method: task_arithmetic
54
- parameters:
55
- normalize: false
56
- slices:
57
- - sources:
58
- - layer_range: [0, 32]
59
- model: grimjim/llama-3-Nephilim-v1-8B
60
- - layer_range: [0, 32]
61
- model: grimjim/Llama-3-Instruct-8B-SPPO-Iter3-SimPO-merge
62
- parameters:
63
- weight: 0.62
64
- - layer_range: [0, 32]
65
- model: openlynn/Llama-3-Soliloquy-8B-v2
66
- parameters:
67
- weight: 0.03
68
- - layer_range: [0, 32]
69
- model: tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1
70
- parameters:
71
- weight: 0.2
72
- - layer_range: [0, 32]
73
- model: MLP-KTLim/llama-3-Korean-Bllossom-8B
74
- parameters:
75
- weight: 0.001
76
- - layer_range: [0, 32]
77
- model: grimjim/llama-3-aaditya-OpenBioLLM-8B
78
- parameters:
79
- weight: 0.03
80
-
81
- ```
 
1
+ ---
2
+ base_model: grimjim/llama-3-Nephilim-v2.1-8B
3
+ library_name: transformers
4
+ quanted_by: grimjim
5
+ license: cc-by-nc-4.0
6
+ pipeline_tag: text-generation
7
+ ---
8
+ # grimjim/llama-3-Nephilim-v2.1-8B-GGUF
9
+
10
+ This repo contains select GGUF quants of a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
11
+
12
+ Will bf16 weights are [available here](https://huggingface.co/grimjim/llama-3-Nephilim-v2.1-8B).
13
+
14
+ Unorthodox contributions to this merge include a biomedical model, a roleplay-focused model, a Japanese language model, and a Korean language model; the original base model was an abliterated model combined with a cybersecurity model. A merger of two recent preference optimization models was also included to shore up reasoning, with the expectation that overall good reasoning and coherence will contribute to narrative flow. Roleplay text generation appears to be easily steered toward emotionally intelligent characters. It should be noted that the vast majority of merge contributions (as measured by weight) were not intended for roleplay.
15
+
16
+ Tested with 8k context length, temperature=1, and minP=0.01 using Instruct prompts for Llama 3. Example Llama 3 presets can be [downloaded here](https://huggingface.co/debased-ai/SillyTavern-settings/tree/main/presets). Specifically, "Llama 3 Lukewarm Presets 8k context.json" corresponds to test settings.
17
+
18
+ Example context templates variants tested with Llama 3 can be [downloaded here](https://huggingface.co/debased-ai/SillyTavern-settings/tree/main/advanced_formatting/context_template); their corresponding Instruct prompts can be [downloaded here](https://huggingface.co/debased-ai/SillyTavern-settings/tree/main/advanced_formatting/instruct_mode). Specifically, the two "Llama 3 Instruct Safety.json" files correspond to test settings.
19
+
20
+ Care should be taken when using this model, as it is possible that harmful outputs may be generated. Given that this model is derivative, responsible use is further mandated by the [WhiteRabbitNeo Usage Restrictions Extension](https://huggingface.co/WhiteRabbitNeo/Llama-3-WhiteRabbitNeo-8B-v2.0) to the Llama-3 License. This model is further subject to CC-BY-NC-4.0 by default, meaning that commercial use is restricted, barring an alternative licensing agreement.
21
+
22
+ Built with Meta Llama 3.
23
+
24
+ ## Merge Details
25
+ ### Merge Method
26
+
27
+ This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [grimjim/llama-3-Nephilim-v1-8B](https://huggingface.co/grimjim/llama-3-Nephilim-v1-8B) as a base.
28
+
29
+ ### Models Merged
30
+
31
+ The following models were included in the merge:
32
+ * [openlynn/Llama-3-Soliloquy-8B-v2](https://huggingface.co/openlynn/Llama-3-Soliloquy-8B-v2)
33
+ * [grimjim/Llama-3-Instruct-8B-SPPO-Iter3-SimPO-merge](https://huggingface.co/grimjim/Llama-3-Instruct-8B-SPPO-Iter3-SimPO-merge)
34
+ * [grimjim/llama-3-aaditya-OpenBioLLM-8B](https://huggingface.co/grimjim/llama-3-aaditya-OpenBioLLM-8B)
35
+ * [tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1](https://huggingface.co/tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1)
36
+ * [MLP-KTLim/llama-3-Korean-Bllossom-8B](https://huggingface.co/MLP-KTLim/llama-3-Korean-Bllossom-8B)
37
+
38
+ ### Configuration
39
+
40
+ The following YAML configuration was used to produce this model:
41
+
42
+ ```yaml
43
+ base_model: grimjim/llama-3-Nephilim-v1-8B
44
+ dtype: bfloat16
45
+ merge_method: task_arithmetic
46
+ parameters:
47
+ normalize: false
48
+ slices:
49
+ - sources:
50
+ - layer_range: [0, 32]
51
+ model: grimjim/llama-3-Nephilim-v1-8B
52
+ - layer_range: [0, 32]
53
+ model: grimjim/Llama-3-Instruct-8B-SPPO-Iter3-SimPO-merge
54
+ parameters:
55
+ weight: 0.62
56
+ - layer_range: [0, 32]
57
+ model: openlynn/Llama-3-Soliloquy-8B-v2
58
+ parameters:
59
+ weight: 0.03
60
+ - layer_range: [0, 32]
61
+ model: tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1
62
+ parameters:
63
+ weight: 0.2
64
+ - layer_range: [0, 32]
65
+ model: MLP-KTLim/llama-3-Korean-Bllossom-8B
66
+ parameters:
67
+ weight: 0.001
68
+ - layer_range: [0, 32]
69
+ model: grimjim/llama-3-aaditya-OpenBioLLM-8B
70
+ parameters:
71
+ weight: 0.03
72
+
73
+ ```