genevera commited on
Commit
3b61a47
1 Parent(s): 642bf55

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +166 -0
README.md ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: AdaptLLM/medicine-chat
3
+ datasets:
4
+ - EleutherAI/pile
5
+ - Open-Orca/OpenOrca
6
+ - GAIR/lima
7
+ - WizardLM/WizardLM_evol_instruct_V2_196k
8
+ language:
9
+ - en
10
+ license: llama2
11
+ metrics:
12
+ - accuracy
13
+ pipeline_tag: text-generation
14
+ tags:
15
+ - biology
16
+ - medical
17
+ - llama-cpp
18
+ - gguf-my-repo
19
+ model-index:
20
+ - name: medicine-chat
21
+ results:
22
+ - task:
23
+ type: text-generation
24
+ name: Text Generation
25
+ dataset:
26
+ name: AI2 Reasoning Challenge (25-Shot)
27
+ type: ai2_arc
28
+ config: ARC-Challenge
29
+ split: test
30
+ args:
31
+ num_few_shot: 25
32
+ metrics:
33
+ - type: acc_norm
34
+ value: 53.75
35
+ name: normalized accuracy
36
+ source:
37
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=AdaptLLM/medicine-chat
38
+ name: Open LLM Leaderboard
39
+ - task:
40
+ type: text-generation
41
+ name: Text Generation
42
+ dataset:
43
+ name: HellaSwag (10-Shot)
44
+ type: hellaswag
45
+ split: validation
46
+ args:
47
+ num_few_shot: 10
48
+ metrics:
49
+ - type: acc_norm
50
+ value: 76.11
51
+ name: normalized accuracy
52
+ source:
53
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=AdaptLLM/medicine-chat
54
+ name: Open LLM Leaderboard
55
+ - task:
56
+ type: text-generation
57
+ name: Text Generation
58
+ dataset:
59
+ name: MMLU (5-Shot)
60
+ type: cais/mmlu
61
+ config: all
62
+ split: test
63
+ args:
64
+ num_few_shot: 5
65
+ metrics:
66
+ - type: acc
67
+ value: 49.98
68
+ name: accuracy
69
+ source:
70
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=AdaptLLM/medicine-chat
71
+ name: Open LLM Leaderboard
72
+ - task:
73
+ type: text-generation
74
+ name: Text Generation
75
+ dataset:
76
+ name: TruthfulQA (0-shot)
77
+ type: truthful_qa
78
+ config: multiple_choice
79
+ split: validation
80
+ args:
81
+ num_few_shot: 0
82
+ metrics:
83
+ - type: mc2
84
+ value: 43.46
85
+ source:
86
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=AdaptLLM/medicine-chat
87
+ name: Open LLM Leaderboard
88
+ - task:
89
+ type: text-generation
90
+ name: Text Generation
91
+ dataset:
92
+ name: Winogrande (5-shot)
93
+ type: winogrande
94
+ config: winogrande_xl
95
+ split: validation
96
+ args:
97
+ num_few_shot: 5
98
+ metrics:
99
+ - type: acc
100
+ value: 75.69
101
+ name: accuracy
102
+ source:
103
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=AdaptLLM/medicine-chat
104
+ name: Open LLM Leaderboard
105
+ - task:
106
+ type: text-generation
107
+ name: Text Generation
108
+ dataset:
109
+ name: GSM8k (5-shot)
110
+ type: gsm8k
111
+ config: main
112
+ split: test
113
+ args:
114
+ num_few_shot: 5
115
+ metrics:
116
+ - type: acc
117
+ value: 18.95
118
+ name: accuracy
119
+ source:
120
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=AdaptLLM/medicine-chat
121
+ name: Open LLM Leaderboard
122
+ ---
123
+
124
+ # genevera/medicine-chat-Q8_0-GGUF
125
+ This model was converted to GGUF format from [`AdaptLLM/medicine-chat`](https://huggingface.co/AdaptLLM/medicine-chat) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
126
+ Refer to the [original model card](https://huggingface.co/AdaptLLM/medicine-chat) for more details on the model.
127
+
128
+ ## Use with llama.cpp
129
+ Install llama.cpp through brew (works on Mac and Linux)
130
+
131
+ ```bash
132
+ brew install llama.cpp
133
+
134
+ ```
135
+ Invoke the llama.cpp server or the CLI.
136
+
137
+ ### CLI:
138
+ ```bash
139
+ llama-cli --hf-repo genevera/medicine-chat-Q8_0-GGUF --hf-file medicine-chat-q8_0.gguf -p "The meaning to life and the universe is"
140
+ ```
141
+
142
+ ### Server:
143
+ ```bash
144
+ llama-server --hf-repo genevera/medicine-chat-Q8_0-GGUF --hf-file medicine-chat-q8_0.gguf -c 2048
145
+ ```
146
+
147
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
148
+
149
+ Step 1: Clone llama.cpp from GitHub.
150
+ ```
151
+ git clone https://github.com/ggerganov/llama.cpp
152
+ ```
153
+
154
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
155
+ ```
156
+ cd llama.cpp && LLAMA_CURL=1 make
157
+ ```
158
+
159
+ Step 3: Run inference through the main binary.
160
+ ```
161
+ ./llama-cli --hf-repo genevera/medicine-chat-Q8_0-GGUF --hf-file medicine-chat-q8_0.gguf -p "The meaning to life and the universe is"
162
+ ```
163
+ or
164
+ ```
165
+ ./llama-server --hf-repo genevera/medicine-chat-Q8_0-GGUF --hf-file medicine-chat-q8_0.gguf -c 2048
166
+ ```