NoQuest commited on
Commit
09147aa
·
1 Parent(s): 94f5005

mistral et adapters sent

Browse files
.DS_Store ADDED
Binary file (10.2 kB). View file
 
loras/adaptersRéaliséPrMistral.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bf976a86a666903f0947131c4a4f0d4b8d92e4c3c69a171e22530fb48a21895
3
+ size 3416406
models/model/.gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
models/model/README.md ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - finetuned
5
+ pipeline_tag: text-generation
6
+ inference: true
7
+ widget:
8
+ - messages:
9
+ - role: user
10
+ content: What is your favorite condiment?
11
+ ---
12
+
13
+ # Model Card for Mistral-7B-Instruct-v0.2
14
+
15
+ ###
16
+
17
+ > [!CAUTION]
18
+ > ⚠️
19
+ > The `transformers` tokenizer might give incorrect results as it has not been tested by the Mistral team. To make sure that your encoding and decoding is correct, please use `mistral_common` as shown below:
20
+
21
+ ## Encode and Decode with `mistral_common`
22
+
23
+ ```py
24
+ from mistral_common.tokens.tokenizers.mistral import MistralTokenizer
25
+ from mistral_common.protocol.instruct.messages import UserMessage
26
+ from mistral_common.protocol.instruct.request import ChatCompletionRequest
27
+
28
+ mistral_models_path = "MISTRAL_MODELS_PATH"
29
+
30
+ tokenizer = MistralTokenizer.v1()
31
+
32
+ completion_request = ChatCompletionRequest(messages=[UserMessage(content="Explain Machine Learning to me in a nutshell.")])
33
+
34
+ tokens = tokenizer.encode_chat_completion(completion_request).tokens
35
+ ```
36
+
37
+ ## Inference with `mistral_inference`
38
+
39
+ ```py
40
+ from mistral_inference.model import Transformer
41
+ from mistral_inference.generate import generate
42
+
43
+ model = Transformer.from_folder(mistral_models_path)
44
+ out_tokens, _ = generate([tokens], model, max_tokens=64, temperature=0.0, eos_id=tokenizer.instruct_tokenizer.tokenizer.eos_id)
45
+
46
+ result = tokenizer.decode(out_tokens[0])
47
+
48
+ print(result)
49
+ ```
50
+
51
+ ## Inference with hugging face `transformers`
52
+
53
+ ```py
54
+ from transformers import AutoModelForCausalLM
55
+
56
+ model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
57
+ model.to("cuda")
58
+
59
+ generated_ids = model.generate(tokens, max_new_tokens=1000, do_sample=True)
60
+
61
+ # decode with mistral tokenizer
62
+ result = tokenizer.decode(generated_ids[0].tolist())
63
+ print(result)
64
+ ```
65
+
66
+ > [!TIP]
67
+ > PRs to correct the `transformers` tokenizer so that it gives 1-to-1 the same results as the `mistral_common` reference implementation are very welcome!
68
+
69
+ ---
70
+
71
+ The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2.
72
+
73
+ Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1
74
+ - 32k context window (vs 8k context in v0.1)
75
+ - Rope-theta = 1e6
76
+ - No Sliding-Window Attention
77
+
78
+ For full details of this model please read our [paper](https://arxiv.org/abs/2310.06825) and [release blog post](https://mistral.ai/news/la-plateforme/).
79
+
80
+ ## Instruction format
81
+
82
+ In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.
83
+
84
+ E.g.
85
+ ```
86
+ text = "<s>[INST] What is your favourite condiment? [/INST]"
87
+ "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> "
88
+ "[INST] Do you have mayonnaise recipes? [/INST]"
89
+ ```
90
+
91
+ This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method:
92
+
93
+ ```python
94
+ from transformers import AutoModelForCausalLM, AutoTokenizer
95
+
96
+ device = "cuda" # the device to load the model onto
97
+
98
+ model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
99
+ tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
100
+
101
+ messages = [
102
+ {"role": "user", "content": "What is your favourite condiment?"},
103
+ {"role": "assistant", "content": "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!"},
104
+ {"role": "user", "content": "Do you have mayonnaise recipes?"}
105
+ ]
106
+
107
+ encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
108
+
109
+ model_inputs = encodeds.to(device)
110
+ model.to(device)
111
+
112
+ generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
113
+ decoded = tokenizer.batch_decode(generated_ids)
114
+ print(decoded[0])
115
+ ```
116
+
117
+ ## Troubleshooting
118
+ - If you see the following error:
119
+ ```
120
+ Traceback (most recent call last):
121
+ File "", line 1, in
122
+ File "/transformers/models/auto/auto_factory.py", line 482, in from_pretrained
123
+ config, kwargs = AutoConfig.from_pretrained(
124
+ File "/transformers/models/auto/configuration_auto.py", line 1022, in from_pretrained
125
+ config_class = CONFIG_MAPPING[config_dict["model_type"]]
126
+ File "/transformers/models/auto/configuration_auto.py", line 723, in getitem
127
+ raise KeyError(key)
128
+ KeyError: 'mistral'
129
+ ```
130
+
131
+ Installing transformers from source should solve the issue
132
+ pip install git+https://github.com/huggingface/transformers
133
+
134
+ This should not be required after transformers-v4.33.4.
135
+
136
+ ## Limitations
137
+
138
+ The Mistral 7B Instruct model is a quick demonstration that the base model can be easily fine-tuned to achieve compelling performance.
139
+ It does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to
140
+ make the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs.
141
+
142
+ ## The Mistral AI Team
143
+
144
+ Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Blanche Savary, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Emma Bou Hanna, Florian Bressand, Gianna Lengyel, Guillaume Bour, Guillaume Lample, Lélio Renard Lavaud, Louis Ternon, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Théophile Gervet, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed.
models/model/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MistralForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 4096,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 14336,
12
+ "max_position_embeddings": 32768,
13
+ "model_type": "mistral",
14
+ "num_attention_heads": 32,
15
+ "num_hidden_layers": 32,
16
+ "num_key_value_heads": 8,
17
+ "rms_norm_eps": 1e-05,
18
+ "rope_theta": 1000000.0,
19
+ "sliding_window": null,
20
+ "tie_word_embeddings": false,
21
+ "torch_dtype": "bfloat16",
22
+ "transformers_version": "4.36.0",
23
+ "use_cache": true,
24
+ "vocab_size": 32000
25
+ }
models/model/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.36.0"
6
+ }
models/model/nohup.out ADDED
@@ -0,0 +1,212 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ tsc-dyn-get: Using source :github (:loaded nil :recorded 0.18.0 :requested 0.18.0)
2
+ tsc-dyn-get: Recorded version already satifies requested -> loading
3
+ tree-sitter-langs: Grammar bundle v0.12.20 was already installed; skipped
4
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
5
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
6
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
7
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
8
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
9
+ Loading /Users/computer/.emacs.d/myssh.el (source)...
10
+ Loading /Users/computer/.emacs.d/myssh.el (source)...done
11
+ Desktop file in use; not loaded.
12
+ Starting Emacs daemon.
13
+ Unable to start the daemon.
14
+ Another instance of Emacs is running the server, either as daemon or interactively.
15
+ You can use emacsclient to connect to that Emacs process.
16
+ Waiting for Emacs...emacsclient: can't find socket; have you started the server?
17
+ emacsclient: To start the server in Emacs, type "M-x server-start".
18
+ emacsclient: No socket or alternate editor. Please use:
19
+
20
+ --socket-name
21
+ --server-file (or environment variable EMACS_SERVER_FILE)
22
+ --alternate-editor (or environment variable ALTERNATE_EDITOR)
23
+ tsc-dyn-get: Using source :github (:loaded nil :recorded 0.18.0 :requested 0.18.0)
24
+ tsc-dyn-get: Recorded version already satifies requested -> loading
25
+ tree-sitter-langs: Grammar bundle v0.12.20 was already installed; skipped
26
+ emacsclient: can't find socket; have you started the server?
27
+ emacsclient: To start the server in Emacs, type "M-x server-start".
28
+ emacsclient: No socket or alternate editor. Please use:
29
+
30
+ --socket-name
31
+ --server-file (or environment variable EMACS_SERVER_FILE)
32
+ --alternate-editor (or environment variable ALTERNATE_EDITOR)
33
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
34
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
35
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
36
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
37
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
38
+ Loading /Users/computer/.emacs.d/myssh.el (source)...
39
+ Loading /Users/computer/.emacs.d/myssh.el (source)...done
40
+ Desktop file in use; not loaded.
41
+ Starting Emacs daemon.
42
+ Package cl is deprecated
43
+ Waiting for Emacs...emacsclient: can't find socket; have you started the server?
44
+ emacsclient: To start the server in Emacs, type "M-x server-start".
45
+ emacsclient: No socket or alternate editor. Please use:
46
+
47
+ --socket-name
48
+ --server-file (or environment variable EMACS_SERVER_FILE)
49
+ --alternate-editor (or environment variable ALTERNATE_EDITOR)
50
+ emacsclient: can't find socket; have you started the server?
51
+ emacsclient: To start the server in Emacs, type "M-x server-start".
52
+ emacsclient: No socket or alternate editor. Please use:
53
+
54
+ --socket-name
55
+ --server-file (or environment variable EMACS_SERVER_FILE)
56
+ --alternate-editor (or environment variable ALTERNATE_EDITOR)
57
+ tsc-dyn-get: Using source :github (:loaded nil :recorded 0.18.0 :requested 0.18.0)
58
+ tsc-dyn-get: Recorded version already satifies requested -> loading
59
+ tree-sitter-langs: Grammar bundle v0.12.20 was already installed; skipped
60
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
61
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
62
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
63
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
64
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
65
+ Loading /Users/computer/.emacs.d/myssh.el (source)...
66
+ Loading /Users/computer/.emacs.d/myssh.el (source)...done
67
+ Desktop file in use; not loaded.
68
+ Starting Emacs daemon.
69
+ Package cl is deprecated
70
+ Waiting for Emacs...
71
+ tsc-dyn-get: Using source :github (:loaded nil :recorded 0.18.0 :requested 0.18.0)
72
+ tsc-dyn-get: Recorded version already satifies requested -> loading
73
+ tree-sitter-langs: Grammar bundle v0.12.20 was already installed; skipped
74
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
75
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
76
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
77
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
78
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
79
+ Loading /Users/computer/.emacs.d/myssh.el (source)...
80
+ Loading /Users/computer/.emacs.d/myssh.el (source)...done
81
+ Desktop file in use; not loaded.
82
+ Starting Emacs daemon.
83
+ Unable to start the daemon.
84
+ Another instance of Emacs is running the server, either as daemon or interactively.
85
+ You can use emacsclient to connect to that Emacs process.
86
+ Waiting for Emacs...tsc-dyn-get: Using source :github (:loaded nil :recorded 0.18.0 :requested 0.18.0)
87
+ tsc-dyn-get: Recorded version already satifies requested -> loading
88
+ tree-sitter-langs: Grammar bundle v0.12.20 was already installed; skipped
89
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
90
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
91
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
92
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
93
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
94
+ Loading /Users/computer/.emacs.d/myssh.el (source)...
95
+ Loading /Users/computer/.emacs.d/myssh.el (source)...done
96
+ Desktop file in use; not loaded.
97
+ Starting Emacs daemon.
98
+ Unable to start the daemon.
99
+ Another instance of Emacs is running the server, either as daemon or interactively.
100
+ You can use emacsclient to connect to that Emacs process.
101
+
102
+
103
+ Waiting for Emacs...
104
+ TabNine process #<process company-tabnine--process> received event "finished
105
+ ".
106
+ Restarting TabNine process.
107
+ TabNine server started.
108
+ Waiting for Emacs...
109
+ Dropping external cursor update event.
110
+ TabNine process #<process company-tabnine--process> received event "finished
111
+ ".
112
+ Restarting TabNine process.
113
+ TabNine server started.
114
+ Waiting for Emacs...Waiting for Emacs...
115
+
116
+ Waiting for Emacs...Waiting for Emacs...
117
+
118
+
119
+ TabNine process #<process company-tabnine--process> received event "finished
120
+ ".
121
+ Restarting TabNine process.
122
+ TabNine server started.
123
+ Dropping external cursor update event.
124
+ Dropping external cursor update event.
125
+ emacsclient: can't find socket; have you started the server?
126
+ emacsclient: To start the server in Emacs, type "M-x server-start".
127
+ emacsclient: No socket or alternate editor. Please use:
128
+
129
+ --socket-name
130
+ --server-file (or environment variable EMACS_SERVER_FILE)
131
+ --alternate-editor (or environment variable ALTERNATE_EDITOR)
132
+ emacsclient: can't find socket; have you started the server?
133
+ emacsclient: To start the server in Emacs, type "M-x server-start".
134
+ emacsclient: No socket or alternate editor. Please use:
135
+
136
+ --socket-name
137
+ --server-file (or environment variable EMACS_SERVER_FILE)
138
+ --alternate-editor (or environment variable ALTERNATE_EDITOR)
139
+ tsc-dyn-get: Using source :github (:loaded nil :recorded 0.18.0 :requested 0.18.0)
140
+ tsc-dyn-get: Recorded version already satifies requested -> loading
141
+ tree-sitter-langs: Grammar bundle v0.12.20 was already installed; skipped
142
+ emacsclient: can't find socket; have you started the server?
143
+ emacsclient: To start the server in Emacs, type "M-x server-start".
144
+ emacsclient: No socket or alternate editor. Please use:
145
+
146
+ --socket-name
147
+ --server-file (or environment variable EMACS_SERVER_FILE)
148
+ --alternate-editor (or environment variable ALTERNATE_EDITOR)
149
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
150
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
151
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
152
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
153
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
154
+ Loading /Users/computer/.emacs.d/myssh.el (source)...
155
+ Loading /Users/computer/.emacs.d/myssh.el (source)...done
156
+ Desktop file in use; not loaded.
157
+ Starting Emacs daemon.
158
+ Package cl is deprecated
159
+ Waiting for Emacs...
160
+ emacsclient: can't find socket; have you started the server?
161
+ emacsclient: To start the server in Emacs, type "M-x server-start".
162
+ emacsclient: No socket or alternate editor. Please use:
163
+
164
+ --socket-name
165
+ --server-file (or environment variable EMACS_SERVER_FILE)
166
+ --alternate-editor (or environment variable ALTERNATE_EDITOR)
167
+ tsc-dyn-get: Using source :github (:loaded nil :recorded 0.18.0 :requested 0.18.0)
168
+ tsc-dyn-get: Recorded version already satifies requested -> loading
169
+ tree-sitter-langs: Grammar bundle v0.12.20 was already installed; skipped
170
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
171
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
172
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
173
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
174
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
175
+ Loading /Users/computer/.emacs.d/myssh.el (source)...
176
+ Loading /Users/computer/.emacs.d/myssh.el (source)...done
177
+ Desktop file in use; not loaded.
178
+ Starting Emacs daemon.
179
+ Package cl is deprecated
180
+ tsc-dyn-get: Using source :github (:loaded nil :recorded 0.18.0 :requested 0.18.0)
181
+ tsc-dyn-get: Recorded version already satifies requested -> loading
182
+ tree-sitter-langs: Grammar bundle v0.12.20 was already installed; skipped
183
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
184
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
185
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
186
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
187
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
188
+ Loading /Users/computer/.emacs.d/myssh.el (source)...
189
+ Loading /Users/computer/.emacs.d/myssh.el (source)...done
190
+ Desktop file in use; not loaded.
191
+ Starting Emacs daemon.
192
+ Unable to start the daemon.
193
+ Another instance of Emacs is running the server, either as daemon or interactively.
194
+ You can use emacsclient to connect to that Emacs process.
195
+ tsc-dyn-get: Using source :github (:loaded nil :recorded 0.18.0 :requested 0.18.0)
196
+ tsc-dyn-get: Recorded version already satifies requested -> loading
197
+ tree-sitter-langs: Grammar bundle v0.12.20 was already installed; skipped
198
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
199
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
200
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
201
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
202
+ ../../../../../../../.emacs.d/elpa/lsp-mode-20230628.1609/lsp-mode.el: Warning: Non-symbol arguments to cl-defgeneric: (process process)
203
+ Loading /Users/computer/.emacs.d/myssh.el (source)...
204
+ Loading /Users/computer/.emacs.d/myssh.el (source)...done
205
+ Desktop file in use; not loaded.
206
+ Starting Emacs daemon.
207
+ Unable to start the daemon.
208
+ Another instance of Emacs is running the server, either as daemon or interactively.
209
+ You can use emacsclient to connect to that Emacs process.
210
+ Waiting for Emacs...
211
+ Dropping external cursor update event.
212
+ Waiting for Emacs...
models/model/pytorch_model-00001-of-00003.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8836f675fe1c4c43f3ff4e93f4cc0e97ef7a13e8c240fb39ad02d37ff303ef5
3
+ size 4943184288
models/model/pytorch_model-00002-of-00003.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58a7ddffb463397de5dbe1f1e2ec1ccf6aae2b549565f83f3ded124e0b4c5069
3
+ size 4999843272
models/model/pytorch_model-00003-of-00003.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75824d68dcf82d02b731b2bdfd3a9711acb7c58b8d566f4c0d3e9efac52f9a21
3
+ size 5064824210
models/model/pytorch_model.bin.index.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 14483464192
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "pytorch_model-00003-of-00003.bin",
7
+ "model.embed_tokens.weight": "pytorch_model-00001-of-00003.bin",
8
+ "model.layers.0.input_layernorm.weight": "pytorch_model-00001-of-00003.bin",
9
+ "model.layers.0.mlp.down_proj.weight": "pytorch_model-00001-of-00003.bin",
10
+ "model.layers.0.mlp.gate_proj.weight": "pytorch_model-00001-of-00003.bin",
11
+ "model.layers.0.mlp.up_proj.weight": "pytorch_model-00001-of-00003.bin",
12
+ "model.layers.0.post_attention_layernorm.weight": "pytorch_model-00001-of-00003.bin",
13
+ "model.layers.0.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
14
+ "model.layers.0.self_attn.o_proj.weight": "pytorch_model-00001-of-00003.bin",
15
+ "model.layers.0.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
16
+ "model.layers.0.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
17
+ "model.layers.1.input_layernorm.weight": "pytorch_model-00001-of-00003.bin",
18
+ "model.layers.1.mlp.down_proj.weight": "pytorch_model-00001-of-00003.bin",
19
+ "model.layers.1.mlp.gate_proj.weight": "pytorch_model-00001-of-00003.bin",
20
+ "model.layers.1.mlp.up_proj.weight": "pytorch_model-00001-of-00003.bin",
21
+ "model.layers.1.post_attention_layernorm.weight": "pytorch_model-00001-of-00003.bin",
22
+ "model.layers.1.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
23
+ "model.layers.1.self_attn.o_proj.weight": "pytorch_model-00001-of-00003.bin",
24
+ "model.layers.1.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
25
+ "model.layers.1.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
26
+ "model.layers.10.input_layernorm.weight": "pytorch_model-00002-of-00003.bin",
27
+ "model.layers.10.mlp.down_proj.weight": "pytorch_model-00002-of-00003.bin",
28
+ "model.layers.10.mlp.gate_proj.weight": "pytorch_model-00001-of-00003.bin",
29
+ "model.layers.10.mlp.up_proj.weight": "pytorch_model-00001-of-00003.bin",
30
+ "model.layers.10.post_attention_layernorm.weight": "pytorch_model-00002-of-00003.bin",
31
+ "model.layers.10.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
32
+ "model.layers.10.self_attn.o_proj.weight": "pytorch_model-00001-of-00003.bin",
33
+ "model.layers.10.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
34
+ "model.layers.10.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
35
+ "model.layers.11.input_layernorm.weight": "pytorch_model-00002-of-00003.bin",
36
+ "model.layers.11.mlp.down_proj.weight": "pytorch_model-00002-of-00003.bin",
37
+ "model.layers.11.mlp.gate_proj.weight": "pytorch_model-00002-of-00003.bin",
38
+ "model.layers.11.mlp.up_proj.weight": "pytorch_model-00002-of-00003.bin",
39
+ "model.layers.11.post_attention_layernorm.weight": "pytorch_model-00002-of-00003.bin",
40
+ "model.layers.11.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
41
+ "model.layers.11.self_attn.o_proj.weight": "pytorch_model-00002-of-00003.bin",
42
+ "model.layers.11.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
43
+ "model.layers.11.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
44
+ "model.layers.12.input_layernorm.weight": "pytorch_model-00002-of-00003.bin",
45
+ "model.layers.12.mlp.down_proj.weight": "pytorch_model-00002-of-00003.bin",
46
+ "model.layers.12.mlp.gate_proj.weight": "pytorch_model-00002-of-00003.bin",
47
+ "model.layers.12.mlp.up_proj.weight": "pytorch_model-00002-of-00003.bin",
48
+ "model.layers.12.post_attention_layernorm.weight": "pytorch_model-00002-of-00003.bin",
49
+ "model.layers.12.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
50
+ "model.layers.12.self_attn.o_proj.weight": "pytorch_model-00002-of-00003.bin",
51
+ "model.layers.12.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
52
+ "model.layers.12.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
53
+ "model.layers.13.input_layernorm.weight": "pytorch_model-00002-of-00003.bin",
54
+ "model.layers.13.mlp.down_proj.weight": "pytorch_model-00002-of-00003.bin",
55
+ "model.layers.13.mlp.gate_proj.weight": "pytorch_model-00002-of-00003.bin",
56
+ "model.layers.13.mlp.up_proj.weight": "pytorch_model-00002-of-00003.bin",
57
+ "model.layers.13.post_attention_layernorm.weight": "pytorch_model-00002-of-00003.bin",
58
+ "model.layers.13.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
59
+ "model.layers.13.self_attn.o_proj.weight": "pytorch_model-00002-of-00003.bin",
60
+ "model.layers.13.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
61
+ "model.layers.13.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
62
+ "model.layers.14.input_layernorm.weight": "pytorch_model-00002-of-00003.bin",
63
+ "model.layers.14.mlp.down_proj.weight": "pytorch_model-00002-of-00003.bin",
64
+ "model.layers.14.mlp.gate_proj.weight": "pytorch_model-00002-of-00003.bin",
65
+ "model.layers.14.mlp.up_proj.weight": "pytorch_model-00002-of-00003.bin",
66
+ "model.layers.14.post_attention_layernorm.weight": "pytorch_model-00002-of-00003.bin",
67
+ "model.layers.14.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
68
+ "model.layers.14.self_attn.o_proj.weight": "pytorch_model-00002-of-00003.bin",
69
+ "model.layers.14.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
70
+ "model.layers.14.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
71
+ "model.layers.15.input_layernorm.weight": "pytorch_model-00002-of-00003.bin",
72
+ "model.layers.15.mlp.down_proj.weight": "pytorch_model-00002-of-00003.bin",
73
+ "model.layers.15.mlp.gate_proj.weight": "pytorch_model-00002-of-00003.bin",
74
+ "model.layers.15.mlp.up_proj.weight": "pytorch_model-00002-of-00003.bin",
75
+ "model.layers.15.post_attention_layernorm.weight": "pytorch_model-00002-of-00003.bin",
76
+ "model.layers.15.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
77
+ "model.layers.15.self_attn.o_proj.weight": "pytorch_model-00002-of-00003.bin",
78
+ "model.layers.15.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
79
+ "model.layers.15.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
80
+ "model.layers.16.input_layernorm.weight": "pytorch_model-00002-of-00003.bin",
81
+ "model.layers.16.mlp.down_proj.weight": "pytorch_model-00002-of-00003.bin",
82
+ "model.layers.16.mlp.gate_proj.weight": "pytorch_model-00002-of-00003.bin",
83
+ "model.layers.16.mlp.up_proj.weight": "pytorch_model-00002-of-00003.bin",
84
+ "model.layers.16.post_attention_layernorm.weight": "pytorch_model-00002-of-00003.bin",
85
+ "model.layers.16.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
86
+ "model.layers.16.self_attn.o_proj.weight": "pytorch_model-00002-of-00003.bin",
87
+ "model.layers.16.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
88
+ "model.layers.16.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
89
+ "model.layers.17.input_layernorm.weight": "pytorch_model-00002-of-00003.bin",
90
+ "model.layers.17.mlp.down_proj.weight": "pytorch_model-00002-of-00003.bin",
91
+ "model.layers.17.mlp.gate_proj.weight": "pytorch_model-00002-of-00003.bin",
92
+ "model.layers.17.mlp.up_proj.weight": "pytorch_model-00002-of-00003.bin",
93
+ "model.layers.17.post_attention_layernorm.weight": "pytorch_model-00002-of-00003.bin",
94
+ "model.layers.17.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
95
+ "model.layers.17.self_attn.o_proj.weight": "pytorch_model-00002-of-00003.bin",
96
+ "model.layers.17.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
97
+ "model.layers.17.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
98
+ "model.layers.18.input_layernorm.weight": "pytorch_model-00002-of-00003.bin",
99
+ "model.layers.18.mlp.down_proj.weight": "pytorch_model-00002-of-00003.bin",
100
+ "model.layers.18.mlp.gate_proj.weight": "pytorch_model-00002-of-00003.bin",
101
+ "model.layers.18.mlp.up_proj.weight": "pytorch_model-00002-of-00003.bin",
102
+ "model.layers.18.post_attention_layernorm.weight": "pytorch_model-00002-of-00003.bin",
103
+ "model.layers.18.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
104
+ "model.layers.18.self_attn.o_proj.weight": "pytorch_model-00002-of-00003.bin",
105
+ "model.layers.18.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
106
+ "model.layers.18.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
107
+ "model.layers.19.input_layernorm.weight": "pytorch_model-00002-of-00003.bin",
108
+ "model.layers.19.mlp.down_proj.weight": "pytorch_model-00002-of-00003.bin",
109
+ "model.layers.19.mlp.gate_proj.weight": "pytorch_model-00002-of-00003.bin",
110
+ "model.layers.19.mlp.up_proj.weight": "pytorch_model-00002-of-00003.bin",
111
+ "model.layers.19.post_attention_layernorm.weight": "pytorch_model-00002-of-00003.bin",
112
+ "model.layers.19.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
113
+ "model.layers.19.self_attn.o_proj.weight": "pytorch_model-00002-of-00003.bin",
114
+ "model.layers.19.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
115
+ "model.layers.19.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
116
+ "model.layers.2.input_layernorm.weight": "pytorch_model-00001-of-00003.bin",
117
+ "model.layers.2.mlp.down_proj.weight": "pytorch_model-00001-of-00003.bin",
118
+ "model.layers.2.mlp.gate_proj.weight": "pytorch_model-00001-of-00003.bin",
119
+ "model.layers.2.mlp.up_proj.weight": "pytorch_model-00001-of-00003.bin",
120
+ "model.layers.2.post_attention_layernorm.weight": "pytorch_model-00001-of-00003.bin",
121
+ "model.layers.2.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
122
+ "model.layers.2.self_attn.o_proj.weight": "pytorch_model-00001-of-00003.bin",
123
+ "model.layers.2.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
124
+ "model.layers.2.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
125
+ "model.layers.20.input_layernorm.weight": "pytorch_model-00002-of-00003.bin",
126
+ "model.layers.20.mlp.down_proj.weight": "pytorch_model-00002-of-00003.bin",
127
+ "model.layers.20.mlp.gate_proj.weight": "pytorch_model-00002-of-00003.bin",
128
+ "model.layers.20.mlp.up_proj.weight": "pytorch_model-00002-of-00003.bin",
129
+ "model.layers.20.post_attention_layernorm.weight": "pytorch_model-00002-of-00003.bin",
130
+ "model.layers.20.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
131
+ "model.layers.20.self_attn.o_proj.weight": "pytorch_model-00002-of-00003.bin",
132
+ "model.layers.20.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
133
+ "model.layers.20.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
134
+ "model.layers.21.input_layernorm.weight": "pytorch_model-00002-of-00003.bin",
135
+ "model.layers.21.mlp.down_proj.weight": "pytorch_model-00002-of-00003.bin",
136
+ "model.layers.21.mlp.gate_proj.weight": "pytorch_model-00002-of-00003.bin",
137
+ "model.layers.21.mlp.up_proj.weight": "pytorch_model-00002-of-00003.bin",
138
+ "model.layers.21.post_attention_layernorm.weight": "pytorch_model-00002-of-00003.bin",
139
+ "model.layers.21.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
140
+ "model.layers.21.self_attn.o_proj.weight": "pytorch_model-00002-of-00003.bin",
141
+ "model.layers.21.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
142
+ "model.layers.21.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
143
+ "model.layers.22.input_layernorm.weight": "pytorch_model-00003-of-00003.bin",
144
+ "model.layers.22.mlp.down_proj.weight": "pytorch_model-00003-of-00003.bin",
145
+ "model.layers.22.mlp.gate_proj.weight": "pytorch_model-00003-of-00003.bin",
146
+ "model.layers.22.mlp.up_proj.weight": "pytorch_model-00003-of-00003.bin",
147
+ "model.layers.22.post_attention_layernorm.weight": "pytorch_model-00003-of-00003.bin",
148
+ "model.layers.22.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
149
+ "model.layers.22.self_attn.o_proj.weight": "pytorch_model-00002-of-00003.bin",
150
+ "model.layers.22.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
151
+ "model.layers.22.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
152
+ "model.layers.23.input_layernorm.weight": "pytorch_model-00003-of-00003.bin",
153
+ "model.layers.23.mlp.down_proj.weight": "pytorch_model-00003-of-00003.bin",
154
+ "model.layers.23.mlp.gate_proj.weight": "pytorch_model-00003-of-00003.bin",
155
+ "model.layers.23.mlp.up_proj.weight": "pytorch_model-00003-of-00003.bin",
156
+ "model.layers.23.post_attention_layernorm.weight": "pytorch_model-00003-of-00003.bin",
157
+ "model.layers.23.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
158
+ "model.layers.23.self_attn.o_proj.weight": "pytorch_model-00003-of-00003.bin",
159
+ "model.layers.23.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
160
+ "model.layers.23.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
161
+ "model.layers.24.input_layernorm.weight": "pytorch_model-00003-of-00003.bin",
162
+ "model.layers.24.mlp.down_proj.weight": "pytorch_model-00003-of-00003.bin",
163
+ "model.layers.24.mlp.gate_proj.weight": "pytorch_model-00003-of-00003.bin",
164
+ "model.layers.24.mlp.up_proj.weight": "pytorch_model-00003-of-00003.bin",
165
+ "model.layers.24.post_attention_layernorm.weight": "pytorch_model-00003-of-00003.bin",
166
+ "model.layers.24.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
167
+ "model.layers.24.self_attn.o_proj.weight": "pytorch_model-00003-of-00003.bin",
168
+ "model.layers.24.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
169
+ "model.layers.24.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
170
+ "model.layers.25.input_layernorm.weight": "pytorch_model-00003-of-00003.bin",
171
+ "model.layers.25.mlp.down_proj.weight": "pytorch_model-00003-of-00003.bin",
172
+ "model.layers.25.mlp.gate_proj.weight": "pytorch_model-00003-of-00003.bin",
173
+ "model.layers.25.mlp.up_proj.weight": "pytorch_model-00003-of-00003.bin",
174
+ "model.layers.25.post_attention_layernorm.weight": "pytorch_model-00003-of-00003.bin",
175
+ "model.layers.25.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
176
+ "model.layers.25.self_attn.o_proj.weight": "pytorch_model-00003-of-00003.bin",
177
+ "model.layers.25.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
178
+ "model.layers.25.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
179
+ "model.layers.26.input_layernorm.weight": "pytorch_model-00003-of-00003.bin",
180
+ "model.layers.26.mlp.down_proj.weight": "pytorch_model-00003-of-00003.bin",
181
+ "model.layers.26.mlp.gate_proj.weight": "pytorch_model-00003-of-00003.bin",
182
+ "model.layers.26.mlp.up_proj.weight": "pytorch_model-00003-of-00003.bin",
183
+ "model.layers.26.post_attention_layernorm.weight": "pytorch_model-00003-of-00003.bin",
184
+ "model.layers.26.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
185
+ "model.layers.26.self_attn.o_proj.weight": "pytorch_model-00003-of-00003.bin",
186
+ "model.layers.26.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
187
+ "model.layers.26.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
188
+ "model.layers.27.input_layernorm.weight": "pytorch_model-00003-of-00003.bin",
189
+ "model.layers.27.mlp.down_proj.weight": "pytorch_model-00003-of-00003.bin",
190
+ "model.layers.27.mlp.gate_proj.weight": "pytorch_model-00003-of-00003.bin",
191
+ "model.layers.27.mlp.up_proj.weight": "pytorch_model-00003-of-00003.bin",
192
+ "model.layers.27.post_attention_layernorm.weight": "pytorch_model-00003-of-00003.bin",
193
+ "model.layers.27.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
194
+ "model.layers.27.self_attn.o_proj.weight": "pytorch_model-00003-of-00003.bin",
195
+ "model.layers.27.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
196
+ "model.layers.27.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
197
+ "model.layers.28.input_layernorm.weight": "pytorch_model-00003-of-00003.bin",
198
+ "model.layers.28.mlp.down_proj.weight": "pytorch_model-00003-of-00003.bin",
199
+ "model.layers.28.mlp.gate_proj.weight": "pytorch_model-00003-of-00003.bin",
200
+ "model.layers.28.mlp.up_proj.weight": "pytorch_model-00003-of-00003.bin",
201
+ "model.layers.28.post_attention_layernorm.weight": "pytorch_model-00003-of-00003.bin",
202
+ "model.layers.28.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
203
+ "model.layers.28.self_attn.o_proj.weight": "pytorch_model-00003-of-00003.bin",
204
+ "model.layers.28.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
205
+ "model.layers.28.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
206
+ "model.layers.29.input_layernorm.weight": "pytorch_model-00003-of-00003.bin",
207
+ "model.layers.29.mlp.down_proj.weight": "pytorch_model-00003-of-00003.bin",
208
+ "model.layers.29.mlp.gate_proj.weight": "pytorch_model-00003-of-00003.bin",
209
+ "model.layers.29.mlp.up_proj.weight": "pytorch_model-00003-of-00003.bin",
210
+ "model.layers.29.post_attention_layernorm.weight": "pytorch_model-00003-of-00003.bin",
211
+ "model.layers.29.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
212
+ "model.layers.29.self_attn.o_proj.weight": "pytorch_model-00003-of-00003.bin",
213
+ "model.layers.29.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
214
+ "model.layers.29.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
215
+ "model.layers.3.input_layernorm.weight": "pytorch_model-00001-of-00003.bin",
216
+ "model.layers.3.mlp.down_proj.weight": "pytorch_model-00001-of-00003.bin",
217
+ "model.layers.3.mlp.gate_proj.weight": "pytorch_model-00001-of-00003.bin",
218
+ "model.layers.3.mlp.up_proj.weight": "pytorch_model-00001-of-00003.bin",
219
+ "model.layers.3.post_attention_layernorm.weight": "pytorch_model-00001-of-00003.bin",
220
+ "model.layers.3.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
221
+ "model.layers.3.self_attn.o_proj.weight": "pytorch_model-00001-of-00003.bin",
222
+ "model.layers.3.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
223
+ "model.layers.3.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
224
+ "model.layers.30.input_layernorm.weight": "pytorch_model-00003-of-00003.bin",
225
+ "model.layers.30.mlp.down_proj.weight": "pytorch_model-00003-of-00003.bin",
226
+ "model.layers.30.mlp.gate_proj.weight": "pytorch_model-00003-of-00003.bin",
227
+ "model.layers.30.mlp.up_proj.weight": "pytorch_model-00003-of-00003.bin",
228
+ "model.layers.30.post_attention_layernorm.weight": "pytorch_model-00003-of-00003.bin",
229
+ "model.layers.30.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
230
+ "model.layers.30.self_attn.o_proj.weight": "pytorch_model-00003-of-00003.bin",
231
+ "model.layers.30.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
232
+ "model.layers.30.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
233
+ "model.layers.31.input_layernorm.weight": "pytorch_model-00003-of-00003.bin",
234
+ "model.layers.31.mlp.down_proj.weight": "pytorch_model-00003-of-00003.bin",
235
+ "model.layers.31.mlp.gate_proj.weight": "pytorch_model-00003-of-00003.bin",
236
+ "model.layers.31.mlp.up_proj.weight": "pytorch_model-00003-of-00003.bin",
237
+ "model.layers.31.post_attention_layernorm.weight": "pytorch_model-00003-of-00003.bin",
238
+ "model.layers.31.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
239
+ "model.layers.31.self_attn.o_proj.weight": "pytorch_model-00003-of-00003.bin",
240
+ "model.layers.31.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
241
+ "model.layers.31.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
242
+ "model.layers.4.input_layernorm.weight": "pytorch_model-00001-of-00003.bin",
243
+ "model.layers.4.mlp.down_proj.weight": "pytorch_model-00001-of-00003.bin",
244
+ "model.layers.4.mlp.gate_proj.weight": "pytorch_model-00001-of-00003.bin",
245
+ "model.layers.4.mlp.up_proj.weight": "pytorch_model-00001-of-00003.bin",
246
+ "model.layers.4.post_attention_layernorm.weight": "pytorch_model-00001-of-00003.bin",
247
+ "model.layers.4.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
248
+ "model.layers.4.self_attn.o_proj.weight": "pytorch_model-00001-of-00003.bin",
249
+ "model.layers.4.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
250
+ "model.layers.4.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
251
+ "model.layers.5.input_layernorm.weight": "pytorch_model-00001-of-00003.bin",
252
+ "model.layers.5.mlp.down_proj.weight": "pytorch_model-00001-of-00003.bin",
253
+ "model.layers.5.mlp.gate_proj.weight": "pytorch_model-00001-of-00003.bin",
254
+ "model.layers.5.mlp.up_proj.weight": "pytorch_model-00001-of-00003.bin",
255
+ "model.layers.5.post_attention_layernorm.weight": "pytorch_model-00001-of-00003.bin",
256
+ "model.layers.5.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
257
+ "model.layers.5.self_attn.o_proj.weight": "pytorch_model-00001-of-00003.bin",
258
+ "model.layers.5.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
259
+ "model.layers.5.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
260
+ "model.layers.6.input_layernorm.weight": "pytorch_model-00001-of-00003.bin",
261
+ "model.layers.6.mlp.down_proj.weight": "pytorch_model-00001-of-00003.bin",
262
+ "model.layers.6.mlp.gate_proj.weight": "pytorch_model-00001-of-00003.bin",
263
+ "model.layers.6.mlp.up_proj.weight": "pytorch_model-00001-of-00003.bin",
264
+ "model.layers.6.post_attention_layernorm.weight": "pytorch_model-00001-of-00003.bin",
265
+ "model.layers.6.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
266
+ "model.layers.6.self_attn.o_proj.weight": "pytorch_model-00001-of-00003.bin",
267
+ "model.layers.6.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
268
+ "model.layers.6.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
269
+ "model.layers.7.input_layernorm.weight": "pytorch_model-00001-of-00003.bin",
270
+ "model.layers.7.mlp.down_proj.weight": "pytorch_model-00001-of-00003.bin",
271
+ "model.layers.7.mlp.gate_proj.weight": "pytorch_model-00001-of-00003.bin",
272
+ "model.layers.7.mlp.up_proj.weight": "pytorch_model-00001-of-00003.bin",
273
+ "model.layers.7.post_attention_layernorm.weight": "pytorch_model-00001-of-00003.bin",
274
+ "model.layers.7.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
275
+ "model.layers.7.self_attn.o_proj.weight": "pytorch_model-00001-of-00003.bin",
276
+ "model.layers.7.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
277
+ "model.layers.7.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
278
+ "model.layers.8.input_layernorm.weight": "pytorch_model-00001-of-00003.bin",
279
+ "model.layers.8.mlp.down_proj.weight": "pytorch_model-00001-of-00003.bin",
280
+ "model.layers.8.mlp.gate_proj.weight": "pytorch_model-00001-of-00003.bin",
281
+ "model.layers.8.mlp.up_proj.weight": "pytorch_model-00001-of-00003.bin",
282
+ "model.layers.8.post_attention_layernorm.weight": "pytorch_model-00001-of-00003.bin",
283
+ "model.layers.8.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
284
+ "model.layers.8.self_attn.o_proj.weight": "pytorch_model-00001-of-00003.bin",
285
+ "model.layers.8.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
286
+ "model.layers.8.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
287
+ "model.layers.9.input_layernorm.weight": "pytorch_model-00001-of-00003.bin",
288
+ "model.layers.9.mlp.down_proj.weight": "pytorch_model-00001-of-00003.bin",
289
+ "model.layers.9.mlp.gate_proj.weight": "pytorch_model-00001-of-00003.bin",
290
+ "model.layers.9.mlp.up_proj.weight": "pytorch_model-00001-of-00003.bin",
291
+ "model.layers.9.post_attention_layernorm.weight": "pytorch_model-00001-of-00003.bin",
292
+ "model.layers.9.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
293
+ "model.layers.9.self_attn.o_proj.weight": "pytorch_model-00001-of-00003.bin",
294
+ "model.layers.9.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
295
+ "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
296
+ "model.norm.weight": "pytorch_model-00003-of-00003.bin"
297
+ }
298
+ }
models/model/special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
models/model/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
models/model/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
models/model/tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "additional_special_tokens": [],
32
+ "bos_token": "<s>",
33
+ "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{{- bos_token }}\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {%- if loop.first and system_message is defined %}\n {{- ' [INST] ' + system_message + '\\n\\n' + message['content'] + ' [/INST]' }}\n {%- else %}\n {{- ' [INST] ' + message['content'] + ' [/INST]' }}\n {%- endif %}\n {%- elif message['role'] == 'assistant' %}\n {{- ' ' + message['content'] + eos_token}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}\n",
34
+ "clean_up_tokenization_spaces": false,
35
+ "eos_token": "</s>",
36
+ "legacy": false,
37
+ "model_max_length": 1000000000000000019884624838656,
38
+ "pad_token": null,
39
+ "sp_model_kwargs": {},
40
+ "spaces_between_special_tokens": false,
41
+ "tokenizer_class": "LlamaTokenizer",
42
+ "unk_token": "<unk>",
43
+ "use_default_system_prompt": false
44
+ }