ruili0 commited on
Commit
f3a2d76
·
verified ·
1 Parent(s): 7e1dc16

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -5,15 +5,16 @@ license: mit
5
 
6
 
7
 
8
- <a href='https://arxiv.org/abs/'><img src='https://img.shields.io/badge/arXiv-paper-red'></a><a href='https://ruili33.github.io/tpo_website.github.io/'><img src='https://img.shields.io/badge/project-TPO-blue'></a><a href='https://huggingface.co/collections/ruili0/temporal-preference-optimization-67874b451f65db189fa35e10'><img src='https://img.shields.io/badge/huggingface-datasets-green'></a>
9
  <a href='https://huggingface.co/collections/ruili0/temporal-preference-optimization-67874b451f65db189fa35e10'><img src='https://img.shields.io/badge/model-checkpoints-yellow'></a>
10
  <a href='https://github.com/ruili33/TPO'><img src='https://img.shields.io/badge/github-repository-purple'></a>
11
  <img src="cvpr_figure_TPO.png"></img>
12
  # LLaVA-Video-7B-Qwen2-TPO
13
 
14
  LLaVA-Video-7B-Qwen2-TPO, introduced by paper [Temporal Preference Optimization for Long-form Video Understanding](https://arxiv.org/abs), optimized
15
- by temporal preference based on LLaVA-Video-7B-Qwen2. LLaVA-Video-7B-Qwen2-TPO achieves state-of-the-art performance on a variety of benchmarks,
16
- achieving the strongest 7B model on the Video-MME benchmark.
 
17
 
18
 
19
 
 
5
 
6
 
7
 
8
+ <a href='https://arxiv.org/abs/'><img src='https://img.shields.io/badge/arXiv-paper-red'></a><a href='https://ruili33.github.io/tpo_website.github.io/'><img src='https://img.shields.io/badge/project-TPO-blue'></a>
9
  <a href='https://huggingface.co/collections/ruili0/temporal-preference-optimization-67874b451f65db189fa35e10'><img src='https://img.shields.io/badge/model-checkpoints-yellow'></a>
10
  <a href='https://github.com/ruili33/TPO'><img src='https://img.shields.io/badge/github-repository-purple'></a>
11
  <img src="cvpr_figure_TPO.png"></img>
12
  # LLaVA-Video-7B-Qwen2-TPO
13
 
14
  LLaVA-Video-7B-Qwen2-TPO, introduced by paper [Temporal Preference Optimization for Long-form Video Understanding](https://arxiv.org/abs), optimized
15
+ by temporal preference based on LLaVA-Video-7B-Qwen2. The LLaVA-Video-7B-Qwen2-TPO model establishes state-of-the-art performance across a range of
16
+ benchmarks, demonstrating an average performance improvement of 1.5% compared to LLaVA-Video-7B.
17
+ Notably, it emerges as the leading 7B parameter model on the Video-MME benchmark.
18
 
19
 
20