--- library_name: transformers model_name: Vikhr-2.5-VL-2b-Instruct-experemental base_model: - Qwen/Qwen2-VL-2B language: - ru - en license: apache-2.0 --- # 💨👁️ Vikhr-2.5-VL-2b-Instruct-experemental **Vikhr-2.5-VL-2b-Instruct-experemental** — это компактная VLM модель, обученная на переведенном датасете **LLAVA-150K**, специально доученная для обработки на русском языке. Дообученная модель является эксперементальной и не всегда будет работать ожидаемо (особенно OCR). Для обратной связи используйте [Vikhr Models](https://t.me/vikhrlabs) ## Попробовать / Try now: [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/18n9_Aylc87EviAgZeQjlGTLGz-FQ2Q4l?usp=sharing) ### Авторы / Authors - Nikolay Kompanets, [LakoMoor](https://t.me/lakomoordev), [Vikhr Team](https://t.me/vikhrlabs) - Sergey Bratchikov, [NlpWonder](https://t.me/nlpwanderer) - Konstantin Korolev, [Vikhr Team](https://t.me/vikhrlabs) - Aleksandr Nikolich, [Vikhr Team](https://t.me/vikhrlabs) ``` @inproceedings{nikolich2024vikhr, title={Vikhr: Constructing a State-of-the-art Bilingual Open-Source Instruction-Following Large Language Model for {Russian}}, author={Aleksandr Nikolich and Konstantin Korolev and Sergei Bratchikov and Nikolay Kompanets and Igor Kiselev and Artem Shelmanov }, booktitle = {Proceedings of the 4rd Workshop on Multilingual Representation Learning (MRL) @ EMNLP-2024} year={2024}, publisher = {Association for Computational Linguistics}, url={https://arxiv.org/pdf/2405.13929} } ```