usharma1 commited on
Commit
eb8d90c
1 Parent(s): ebde0d8

Remove PR message

Browse files

transformers has been updated to 4.46.0 which contains the necessary components for Idefics3. Hence, the PR message is not needed.

Files changed (1) hide show
  1. README.md +0 -2
README.md CHANGED
@@ -18,8 +18,6 @@ library_name: transformers
18
  <img src="https://huggingface.co/HuggingFaceM4/idefics-80b/resolve/main/assets/IDEFICS.png" alt="Idefics-Obelics logo" width="200" height="100">
19
  </p>
20
 
21
- **Transformers version**: until the next Transformers pypi release, please install Transformers from source and use [this PR](https://github.com/huggingface/transformers/pull/32473) to be able to use Idefics3. TODO: change when new version.
22
-
23
  # Idefics3
24
 
25
  Idefics3 is an open multimodal model that accepts arbitrary sequences of image and text inputs and produces text outputs. The model can answer questions about images, describe visual content, create stories grounded on multiple images, or simply behave as a pure language model without visual inputs. It improves upon [Idefics1](https://huggingface.co/HuggingFaceM4/idefics-80b-instruct) and [Idefics2](https://huggingface.co/HuggingFaceM4/idefics2-8b), significantly enhancing capabilities around OCR, document understanding and visual reasoning.
 
18
  <img src="https://huggingface.co/HuggingFaceM4/idefics-80b/resolve/main/assets/IDEFICS.png" alt="Idefics-Obelics logo" width="200" height="100">
19
  </p>
20
 
 
 
21
  # Idefics3
22
 
23
  Idefics3 is an open multimodal model that accepts arbitrary sequences of image and text inputs and produces text outputs. The model can answer questions about images, describe visual content, create stories grounded on multiple images, or simply behave as a pure language model without visual inputs. It improves upon [Idefics1](https://huggingface.co/HuggingFaceM4/idefics-80b-instruct) and [Idefics2](https://huggingface.co/HuggingFaceM4/idefics2-8b), significantly enhancing capabilities around OCR, document understanding and visual reasoning.