automerger commited on
Commit
344e0c2
1 Parent(s): 77d436c

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,39 +1,29 @@
1
  ---
2
- base_model:
3
- - Gille/StrangeMerges_32-7B-slerp
4
- - mayacinka/yam-jom-7B-slerp
5
- library_name: transformers
6
  tags:
7
- - mergekit
8
  - merge
9
-
 
 
 
 
 
10
  ---
11
- # merge
12
-
13
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
14
-
15
- ## Merge Details
16
- ### Merge Method
17
 
18
- This model was merged using the SLERP merge method.
19
 
20
- ### Models Merged
21
-
22
- The following models were included in the merge:
23
  * [Gille/StrangeMerges_32-7B-slerp](https://huggingface.co/Gille/StrangeMerges_32-7B-slerp)
24
- * [mayacinka/yam-jom-7B-slerp](https://huggingface.co/mayacinka/yam-jom-7B-slerp)
25
-
26
- ### Configuration
27
 
28
- The following YAML configuration was used to produce this model:
29
 
30
  ```yaml
31
-
32
  slices:
33
  - sources:
34
  - model: Gille/StrangeMerges_32-7B-slerp
35
  layer_range: [0, 32]
36
- - model: mayacinka/yam-jom-7B-slerp
37
  layer_range: [0, 32]
38
  merge_method: slerp
39
  base_model: Gille/StrangeMerges_32-7B-slerp
@@ -46,5 +36,29 @@ parameters:
46
  - value: 0.5
47
  dtype: bfloat16
48
  random_seed: 0
49
-
50
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: apache-2.0
 
 
 
3
  tags:
 
4
  - merge
5
+ - mergekit
6
+ - lazymergekit
7
+ - automerger
8
+ base_model:
9
+ - Gille/StrangeMerges_32-7B-slerp
10
+ - mayacinka/yam-jom-7B
11
  ---
 
 
 
 
 
 
12
 
13
+ # Strangemerges_32Yam-7B
14
 
15
+ Strangemerges_32Yam-7B is an automated merge created by [Maxime Labonne](https://huggingface.co/mlabonne) using the following configuration.
 
 
16
  * [Gille/StrangeMerges_32-7B-slerp](https://huggingface.co/Gille/StrangeMerges_32-7B-slerp)
17
+ * [mayacinka/yam-jom-7B](https://huggingface.co/mayacinka/yam-jom-7B)
 
 
18
 
19
+ ## 🧩 Configuration
20
 
21
  ```yaml
 
22
  slices:
23
  - sources:
24
  - model: Gille/StrangeMerges_32-7B-slerp
25
  layer_range: [0, 32]
26
+ - model: mayacinka/yam-jom-7B
27
  layer_range: [0, 32]
28
  merge_method: slerp
29
  base_model: Gille/StrangeMerges_32-7B-slerp
 
36
  - value: 0.5
37
  dtype: bfloat16
38
  random_seed: 0
39
+ ```
40
+
41
+ ## 💻 Usage
42
+
43
+ ```python
44
+ !pip install -qU transformers accelerate
45
+
46
+ from transformers import AutoTokenizer
47
+ import transformers
48
+ import torch
49
+
50
+ model = "automerger/Strangemerges_32Yam-7B"
51
+ messages = [{"role": "user", "content": "What is a large language model?"}]
52
+
53
+ tokenizer = AutoTokenizer.from_pretrained(model)
54
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
55
+ pipeline = transformers.pipeline(
56
+ "text-generation",
57
+ model=model,
58
+ torch_dtype=torch.float16,
59
+ device_map="auto",
60
+ )
61
+
62
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
63
+ print(outputs[0]["generated_text"])
64
+ ```
mergekit_config.yml CHANGED
@@ -3,7 +3,7 @@ slices:
3
  - sources:
4
  - model: Gille/StrangeMerges_32-7B-slerp
5
  layer_range: [0, 32]
6
- - model: mayacinka/yam-jom-7B-slerp
7
  layer_range: [0, 32]
8
  merge_method: slerp
9
  base_model: Gille/StrangeMerges_32-7B-slerp
 
3
  - sources:
4
  - model: Gille/StrangeMerges_32-7B-slerp
5
  layer_range: [0, 32]
6
+ - model: mayacinka/yam-jom-7B
7
  layer_range: [0, 32]
8
  merge_method: slerp
9
  base_model: Gille/StrangeMerges_32-7B-slerp
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a756520e46b3d37b3c753b56844bad4ccc5189f267c1045974944dba7e75897b
3
  size 9942981696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae4cfafa3fab50fe507885488bab40a733653d849ae36ee7b5ff65baf7f601a7
3
  size 9942981696
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:22d4ef8911e8e9f4e6f687418eacfe8b5d8e9f962b0f31659513a62929d837ca
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf3c90f319a67f90bf77f3b53dddbe9626f5ac666792421d314225f84d5e1327
3
  size 4540516344