metadata
pipeline_tag: conversational
tags:
- vicuna
- llama
- text-generation-inference
Converted for use with llama.cpp
- 4-bit quantized
- Needs ~10GB of CPU RAM
pipeline_tag: conversational
tags:
- vicuna
- llama
- text-generation-inference