|
--- |
|
library_name: transformers |
|
model_name: Vikhr-2-VL-2b-Instruct-experimental |
|
base_model: |
|
- Qwen/Qwen2-VL-2B-Instruct |
|
language: |
|
- ru |
|
- en |
|
license: apache-2.0 |
|
--- |
|
|
|
# 💨👁️ Vikhr-2-VL-2b-Instruct-experimental |
|
|
|
**Vikhr-2-VL-2b-Instruct-experimental** — это компактная VLM модель на базе [Qwen2-VL-2B-Instruct](https://huggingface.co/Qwen/Qwen2-VL-2B-Instruct), обученная на переведенном датасете **LLAVA-150K**, специально доученная для обработки на русском языке. |
|
Дообученная модель является эксперементальной и не всегда будет работать ожидаемо (особенно OCR). Для обратной связи используйте [Vikhr Models](https://t.me/vikhrlabs) |
|
|
|
## Попробовать / Try now: |
|
|
|
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/18n9_Aylc87EviAgZeQjlGTLGz-FQ2Q4l?usp=sharing) |
|
|
|
|
|
### Авторы / Authors |
|
|
|
- Nikolay Kompanets, [LakoMoor](https://t.me/lakomoordev), [Vikhr Team](https://t.me/vikhrlabs) |
|
- Sergey Bratchikov, [NlpWonder](https://t.me/nlpwanderer), [Vikhr Team](https://t.me/vikhrlabs) |
|
- Konstantin Korolev, [underground](https://t.me/mlunderground), [Vikhr Team](https://t.me/vikhrlabs) |
|
- Aleksandr Nikolich, [Vikhr Team](https://t.me/vikhrlabs) |
|
|
|
``` |
|
@inproceedings{nikolich2024vikhr, |
|
title={Vikhr: Constructing a State-of-the-art Bilingual Open-Source Instruction-Following Large Language Model for {Russian}}, |
|
author={Aleksandr Nikolich and Konstantin Korolev and Sergei Bratchikov and Nikolay Kompanets and Igor Kiselev and Artem Shelmanov }, |
|
booktitle = {Proceedings of the 4rd Workshop on Multilingual Representation Learning (MRL) @ EMNLP-2024} |
|
year={2024}, |
|
publisher = {Association for Computational Linguistics}, |
|
url={https://arxiv.org/pdf/2405.13929} |
|
} |
|
``` |