File size: 1,876 Bytes
e1bbe78 1b47310 da3bbe8 1b47310 da3bbe8 1b47310 da3bbe8 1b47310 da3bbe8 1b47310 da3bbe8 1b47310 da3bbe8 1b47310 da3bbe8 1b47310 da3bbe8 1b47310 da3bbe8 1b47310 da3bbe8 1b47310 e1bbe78 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 |
---
license: apache-2.0
language:
- en
base_model:
- deepseek-ai/deepseek-vl-1.3b-chat
pipeline_tag: image-to-text
---
# Deepseek-VL-1.3b-chat-4bit
![Deepseek Logo](https://cdn.deepseek.com/logo.png)
## Overview
**Deepseek-VL-1.3b-chat-4bit** is a state-of-the-art multimodal model that combines visual and linguistic processing capabilities. It has been optimized for efficient performance by quantizing the model to 4 bits, significantly reducing its size while maintaining high performance.
### Model Details
- **Model Type**: Multimodal Causal Language Model
- **Base Model Size**: 1.3 billion parameters
- **Quantized Size**: Approximately **1.72 GB** (from the original size)
- **Files Included**:
- `config.json`: Model configuration file.
- `model.safetensors`: The quantized model weights.
- `preprocessor_config.json`: Configuration for the preprocessor.
- `processor_config.json`: Configuration for the processor.
- `special_tokens_map.json`: Mapping for special tokens used in the tokenizer.
- `tokenizer.json`: Tokenizer configuration.
- `tokenizer_config.json`: Additional tokenizer settings.
## Quantization
Quantization is a technique used to reduce the model size and improve inference speed by using lower precision arithmetic. In this case, the model was quantized to 4 bits, which means it utilizes 4 bits to represent each weight instead of the typical 16 or 32 bits. This results in:
- **Size Reduction**: The model size has been reduced from several gigabytes to approximately 1.72 GB.
- **Performance**: The quantized model maintains a high level of accuracy and efficiency, making it suitable for deployment in environments with limited resources.
## Installation
To use the **Deepseek-VL-1.3b-chat-4bit** model, follow these steps:
1. **Install the Required Libraries**:
```bash
pip install transformers huggingface-hub |