Update README.md
Browse files
README.md
CHANGED
@@ -8,13 +8,13 @@ library_name: transformers
|
|
8 |
---
|
9 |
|
10 |
|
11 |
-
<a href='https://arxiv.org/abs/'><img src='https://img.shields.io/badge/arXiv-paper-red'></a><a href='https://ruili33.github.io/tpo_website/'><img src='https://img.shields.io/badge/project-TPO-blue'></a><a href='https://huggingface.co/collections/ruili0/temporal-preference-optimization-67874b451f65db189fa35e10'><img src='https://img.shields.io/badge/huggingface-datasets-green'></a>
|
12 |
<a href='https://huggingface.co/collections/ruili0/temporal-preference-optimization-67874b451f65db189fa35e10'><img src='https://img.shields.io/badge/model-checkpoints-yellow'></a>
|
13 |
<a href='https://github.com/ruili33/TPO'><img src='https://img.shields.io/badge/github-repository-purple'></a>
|
14 |
<img src="cvpr_figure_TPO.png"></img>
|
15 |
# LongVA-7B-TPO
|
16 |
|
17 |
-
LongVA-7B-TPO, introduced by paper [Temporal Preference Optimization for Long-form Video Understanding](https://arxiv.org/abs), optimized
|
18 |
by temporal preference based on LongVA-7B. The LongVA-7B-TPO model establishes state-of-the-art performance across a range of
|
19 |
benchmarks, demonstrating an average performance improvement of 2% compared to LongVA-7B.
|
20 |
|
@@ -82,13 +82,22 @@ print(outputs)
|
|
82 |
This project utilizes certain datasets and checkpoints that are subject to their respective original licenses. Users must comply with all terms and conditions of these original licenses, including but not limited to the OpenAI Terms of Use for the dataset and the specific licenses for base language models (Qwen2 license). This project does not impose any additional constraints beyond those stipulated in the original licenses. Furthermore, users are reminded to ensure that their use of the dataset and checkpoints is in compliance with all applicable laws and regulations.
|
83 |
|
84 |
|
85 |
-
## Citation
|
86 |
|
87 |
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
88 |
|
89 |
**BibTeX:**
|
90 |
-
|
91 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
92 |
|
93 |
**References:**
|
94 |
|
|
|
8 |
---
|
9 |
|
10 |
|
11 |
+
<a href='https://arxiv.org/abs/2501.13919v1'><img src='https://img.shields.io/badge/arXiv-paper-red'></a><a href='https://ruili33.github.io/tpo_website/'><img src='https://img.shields.io/badge/project-TPO-blue'></a><a href='https://huggingface.co/collections/ruili0/temporal-preference-optimization-67874b451f65db189fa35e10'><img src='https://img.shields.io/badge/huggingface-datasets-green'></a>
|
12 |
<a href='https://huggingface.co/collections/ruili0/temporal-preference-optimization-67874b451f65db189fa35e10'><img src='https://img.shields.io/badge/model-checkpoints-yellow'></a>
|
13 |
<a href='https://github.com/ruili33/TPO'><img src='https://img.shields.io/badge/github-repository-purple'></a>
|
14 |
<img src="cvpr_figure_TPO.png"></img>
|
15 |
# LongVA-7B-TPO
|
16 |
|
17 |
+
LongVA-7B-TPO, introduced by paper [Temporal Preference Optimization for Long-form Video Understanding](https://arxiv.org/abs/2501.13919v1), optimized
|
18 |
by temporal preference based on LongVA-7B. The LongVA-7B-TPO model establishes state-of-the-art performance across a range of
|
19 |
benchmarks, demonstrating an average performance improvement of 2% compared to LongVA-7B.
|
20 |
|
|
|
82 |
This project utilizes certain datasets and checkpoints that are subject to their respective original licenses. Users must comply with all terms and conditions of these original licenses, including but not limited to the OpenAI Terms of Use for the dataset and the specific licenses for base language models (Qwen2 license). This project does not impose any additional constraints beyond those stipulated in the original licenses. Furthermore, users are reminded to ensure that their use of the dataset and checkpoints is in compliance with all applicable laws and regulations.
|
83 |
|
84 |
|
85 |
+
## Citation
|
86 |
|
87 |
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
88 |
|
89 |
**BibTeX:**
|
90 |
+
```
|
91 |
+
@misc{li2025temporalpreferenceoptimizationlongform,
|
92 |
+
title={Temporal Preference Optimization for Long-Form Video Understanding},
|
93 |
+
author={Rui Li and Xiaohan Wang and Yuhui Zhang and Zeyu Wang and Serena Yeung-Levy},
|
94 |
+
year={2025},
|
95 |
+
eprint={2501.13919},
|
96 |
+
archivePrefix={arXiv},
|
97 |
+
primaryClass={cs.CV},
|
98 |
+
url={https://arxiv.org/abs/2501.13919},
|
99 |
+
}
|
100 |
+
```
|
101 |
|
102 |
**References:**
|
103 |
|