|
--- |
|
language: |
|
- en |
|
license: apache-2.0 |
|
tags: |
|
- text-generation |
|
- mlx |
|
- 4-bit |
|
datasets: |
|
- ehartford/wizard_vicuna_70k_unfiltered |
|
- totally-not-an-llm/EverythingLM-data-V3 |
|
- Open-Orca/SlimOrca-Dedup |
|
- databricks/databricks-dolly-15k |
|
- THUDM/webglm-qa |
|
base_model: JackFram/llama-160m |
|
--- |
|
|
|
# Llama-160M-Chat-v1-4bit-mlx |
|
This model was converted to MLX format from [`Felladrin/Llama-160M-Chat-v1`](https://huggingface.co/Felladrin/Llama-160M-Chat-v1). |
|
Refer to the [original model card](https://huggingface.co/Felladrin/Llama-160M-Chat-v1) for more details on the model. |
|
|
|
## Use with mlx |
|
|
|
```bash |
|
pip install mlx |
|
git clone https://github.com/ml-explore/mlx-examples.git |
|
cd mlx-examples/llms/hf_llm |
|
python generate.py --model mlx-community/Llama-160M-Chat-v1-4bit-mlx --prompt "<|im_start|>system\nYou are a helpful assistant who answers user's questions with details and curiosity.<|im_end|>\n<|im_start|>user\nWhat are some potential applications for quantum computing?<|im_end|>\n<|im_start|>assistant" |
|
``` |
|
|