alaaelnouby nielsr HF staff commited on
Commit
71ceef4
·
verified ·
1 Parent(s): 0410f31

Add link to paper (#1)

Browse files

- Add link to paper (55002eff8fa94ec0ab1ad2f0a8af40cd933c9680)


Co-authored-by: Niels Rogge <[email protected]>

Files changed (1) hide show
  1. README.md +9 -7
README.md CHANGED
@@ -168,7 +168,7 @@ tags:
168
  - pytorch
169
  ---
170
  # Introduction
171
- [[`AIMv2 Paper`](#)] [[`BibTeX`](#citation)]
172
 
173
  We introduce the AIMv2 family of vision models pre-trained with a multimodal autoregressive objective.
174
  AIMv2 pre-training is simple and straightforward to train and scale effectively. Some AIMv2 highlights include:
@@ -226,12 +226,14 @@ outputs = model(**inputs)
226
  ## Citation
227
  If you find our work useful, please consider citing us as:
228
  ```bibtex
229
- @misc{fini2024multimodal,
230
- title = {Multimodal Autoregressive Pre-training of Large Vision Encoders},
231
- author = {Enrico Fini and Mustafa Shukor and Xiujun Li and Philipp Dufter and Michal Klein and David Haldimann and Sai Aitharaju and Victor Guilherme Turrisi da Costa and Louis Béthune and Zhe Gan and Alexander T Toshev and Marcin Eichner and Moin Nabi and Yinfei Yang and Joshua M. Susskind and Alaaeldin El-Nouby},
232
- year = {2024},
233
- archivePrefix = {arXiv},
234
- primaryClass = {cs.CV},
 
 
235
  }
236
  ```
237
 
 
168
  - pytorch
169
  ---
170
  # Introduction
171
+ [[`AIMv2 Paper`](https://arxiv.org/abs/2411.14402)] [[`BibTeX`](#citation)]
172
 
173
  We introduce the AIMv2 family of vision models pre-trained with a multimodal autoregressive objective.
174
  AIMv2 pre-training is simple and straightforward to train and scale effectively. Some AIMv2 highlights include:
 
226
  ## Citation
227
  If you find our work useful, please consider citing us as:
228
  ```bibtex
229
+ @misc{fini2024multimodalautoregressivepretraininglarge,
230
+ title={Multimodal Autoregressive Pre-training of Large Vision Encoders},
231
+ author={Enrico Fini and Mustafa Shukor and Xiujun Li and Philipp Dufter and Michal Klein and David Haldimann and Sai Aitharaju and Victor Guilherme Turrisi da Costa and Louis Béthune and Zhe Gan and Alexander T Toshev and Marcin Eichner and Moin Nabi and Yinfei Yang and Joshua M. Susskind and Alaaeldin El-Nouby},
232
+ year={2024},
233
+ eprint={2411.14402},
234
+ archivePrefix={arXiv},
235
+ primaryClass={cs.CV},
236
+ url={https://arxiv.org/abs/2411.14402},
237
  }
238
  ```
239