Update README.md
Browse files
README.md
CHANGED
@@ -36,7 +36,33 @@ tags:
|
|
36 |
|
37 |
## Run with LlamaEdge
|
38 |
|
39 |
-
- LlamaEdge version: coming soon
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
40 |
|
41 |
## Quantized GGUF Models
|
42 |
|
|
|
36 |
|
37 |
## Run with LlamaEdge
|
38 |
|
39 |
+
- LlamaEdge version: v0.2.9 (coming soon)
|
40 |
+
|
41 |
+
- Prompt template
|
42 |
+
|
43 |
+
- Prompt type: `stablelm-zephyr`
|
44 |
+
|
45 |
+
- Prompt string
|
46 |
+
|
47 |
+
```text
|
48 |
+
<|user|>
|
49 |
+
{prompt}<|endoftext|>
|
50 |
+
<|assistant|>
|
51 |
+
```
|
52 |
+
|
53 |
+
- Reverse prompt: `<|endoftext|>`
|
54 |
+
|
55 |
+
- Run as LlamaEdge service
|
56 |
+
|
57 |
+
```bash
|
58 |
+
wasmedge --dir .:. --nn-preload default:GGML:AUTO:stablelm-2-zephyr-1_6b-Q5_K_M.gguf llama-api-server.wasm -p stablelm-zephyr -r '<|endoftext|>' -c 1024
|
59 |
+
```
|
60 |
+
|
61 |
+
- Run as LlamaEdge command app
|
62 |
+
|
63 |
+
```bash
|
64 |
+
wasmedge --dir .:. --nn-preload default:GGML:AUTO:stablelm-2-zephyr-1_6b-Q5_K_M.gguf llama-chat.wasm -p stablelm-zephyr -r '<|endoftext|>' --temp 0.5 -c 1024
|
65 |
+
```
|
66 |
|
67 |
## Quantized GGUF Models
|
68 |
|