File size: 1,876 Bytes
e1bbe78
 
 
 
 
 
 
 
1b47310
da3bbe8
1b47310
da3bbe8
1b47310
da3bbe8
1b47310
da3bbe8
1b47310
 
 
 
 
 
 
 
 
 
 
 
da3bbe8
1b47310
da3bbe8
1b47310
da3bbe8
1b47310
 
da3bbe8
1b47310
da3bbe8
1b47310
da3bbe8
1b47310
 
e1bbe78
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
---
license: apache-2.0
language:
- en
base_model:
- deepseek-ai/deepseek-vl-1.3b-chat
pipeline_tag: image-to-text
---
# Deepseek-VL-1.3b-chat-4bit

![Deepseek Logo](https://cdn.deepseek.com/logo.png)

## Overview

**Deepseek-VL-1.3b-chat-4bit** is a state-of-the-art multimodal model that combines visual and linguistic processing capabilities. It has been optimized for efficient performance by quantizing the model to 4 bits, significantly reducing its size while maintaining high performance. 

### Model Details
- **Model Type**: Multimodal Causal Language Model
- **Base Model Size**: 1.3 billion parameters
- **Quantized Size**: Approximately **1.72 GB** (from the original size)
- **Files Included**:
  - `config.json`: Model configuration file.
  - `model.safetensors`: The quantized model weights.
  - `preprocessor_config.json`: Configuration for the preprocessor.
  - `processor_config.json`: Configuration for the processor.
  - `special_tokens_map.json`: Mapping for special tokens used in the tokenizer.
  - `tokenizer.json`: Tokenizer configuration.
  - `tokenizer_config.json`: Additional tokenizer settings.

## Quantization

Quantization is a technique used to reduce the model size and improve inference speed by using lower precision arithmetic. In this case, the model was quantized to 4 bits, which means it utilizes 4 bits to represent each weight instead of the typical 16 or 32 bits. This results in:

- **Size Reduction**: The model size has been reduced from several gigabytes to approximately 1.72 GB.
- **Performance**: The quantized model maintains a high level of accuracy and efficiency, making it suitable for deployment in environments with limited resources.

## Installation

To use the **Deepseek-VL-1.3b-chat-4bit** model, follow these steps:

1. **Install the Required Libraries**:
   ```bash
   pip install transformers huggingface-hub