LakoMoor's picture
Update README.md
dd06431 verified
---
library_name: transformers
model_name: Vikhr-2-VL-2b-Instruct-experimental
base_model:
- Qwen/Qwen2-VL-2B-Instruct
language:
- ru
- en
license: apache-2.0
---
# 💨👁️ Vikhr-2-VL-2b-Instruct-experimental
**Vikhr-2-VL-2b-Instruct-experimental** — это компактная VLM модель на базе [Qwen2-VL-2B-Instruct](https://huggingface.co/Qwen/Qwen2-VL-2B-Instruct), обученная на переведенном датасете **LLAVA-150K**, специально доученная для обработки на русском языке.
Дообученная модель является эксперементальной и не всегда будет работать ожидаемо (особенно OCR). Для обратной связи используйте [Vikhr Models](https://t.me/vikhrlabs)
## Попробовать / Try now:
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/18n9_Aylc87EviAgZeQjlGTLGz-FQ2Q4l?usp=sharing)
### Авторы / Authors
- Nikolay Kompanets, [LakoMoor](https://t.me/lakomoordev), [Vikhr Team](https://t.me/vikhrlabs)
- Sergey Bratchikov, [NlpWonder](https://t.me/nlpwanderer), [Vikhr Team](https://t.me/vikhrlabs)
- Konstantin Korolev, [underground](https://t.me/mlunderground), [Vikhr Team](https://t.me/vikhrlabs)
- Aleksandr Nikolich, [Vikhr Team](https://t.me/vikhrlabs)
```
@inproceedings{nikolich2024vikhr,
title={Vikhr: Constructing a State-of-the-art Bilingual Open-Source Instruction-Following Large Language Model for {Russian}},
author={Aleksandr Nikolich and Konstantin Korolev and Sergei Bratchikov and Nikolay Kompanets and Igor Kiselev and Artem Shelmanov },
booktitle = {Proceedings of the 4rd Workshop on Multilingual Representation Learning (MRL) @ EMNLP-2024}
year={2024},
publisher = {Association for Computational Linguistics},
url={https://arxiv.org/pdf/2405.13929}
}
```