Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,16 @@
|
|
1 |
-
---
|
2 |
-
license: other
|
3 |
-
license_name: qwen
|
4 |
-
license_link: https://huggingface.co/Qwen/QVQ-72B-Preview/blob/main/LICENSE
|
5 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: other
|
3 |
+
license_name: qwen
|
4 |
+
license_link: https://huggingface.co/Qwen/QVQ-72B-Preview/blob/main/LICENSE
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
pipeline_tag: image-text-to-text
|
8 |
+
base_model: Qwen/Qwen2-VL-72B
|
9 |
+
tags:
|
10 |
+
- chat
|
11 |
+
- awq
|
12 |
+
library_name: transformers
|
13 |
+
---
|
14 |
+
# QVQ-72B-Preview AWQ 4-Bit Quantized Version
|
15 |
+
|
16 |
+
This repository provides the AWQ 4-bit quantized version of the QVQ-72B-Preview model, originally developed by Qwen. This model's weights are padded with zeros before quantization to ensure compatibility with multi-GPU tensor parallelism by resolving divisibility constraints. The padding minimally impacts computation while enabling efficient scaling across multiple GPUs.
|