Safetensors
etchat_phi3
custom_code
yeliudev commited on
Commit
9c0ecec
·
verified ·
1 Parent(s): beed8cd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -3
README.md CHANGED
@@ -1,3 +1,42 @@
1
- ---
2
- license: bsd-3-clause
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: bsd-3-clause
3
+ ---
4
+
5
+ # E.T. Chat
6
+
7
+ [arXiv](https://arxiv.org/abs/2409.18111) | [Project Page](https://polyu-chenlab.github.io/etbench) | [GitHub](https://github.com/PolyU-ChenLab/ETBench)
8
+
9
+ E.T. Chat is a novel time-sensitive Video-LLM that reformulates timestamp prediction as an embedding matching problem, serving as a strong baseline on E.T. Bench. E.T. Chat consists of a visual encoder, a frame compressor, and a LLM. A special token \<vid\> is introduced to trigger frame embedding matching for timestamp prediction.
10
+
11
+ > [!NOTE]
12
+ > This checkpoint was trained from a mixture of stage-2 and stage-3 data, yielding much better general chatting capabilities but slightly sub-optimal grounding performance. It shall be considered as the default setting for this model.
13
+
14
+ ## 🔖 Model Details
15
+
16
+ ### Model Description
17
+
18
+ - **Developed by:** Ye Liu
19
+ - **Model type:** Multi-modal Large Language Model
20
+ - **Language(s):** English
21
+ - **License:** BSD-3-Clause
22
+
23
+ ### Training Data
24
+
25
+ The stage-2+3 checkpoint of E.T. Chat was trained from [ET-Instruct-164K](https://huggingface.co/datasets/PolyU-ChenLab/ET-Instruct-164K), [VideoChatGPT](https://mbzuaiac-my.sharepoint.com/:f:/g/personal/hanoona_bangalath_mbzuai_ac_ae/EnLRDehrr8lGqHpC5w1zZ9QBnsiVffYy5vCv8Hl14deRcg?e=Ul5DUE), and [LLaVA-1.5-Instruct](https://github.com/haotian-liu/LLaVA?tab=readme-ov-file#visual-instruction-tuning) datasets.
26
+
27
+ ### More Details
28
+
29
+ Please refer to our [GitHub Repository](https://github.com/PolyU-ChenLab/ETBench) for more details about this model.
30
+
31
+ ## 📖 Citation
32
+
33
+ Please kindly cite our paper if you find this project helpful.
34
+
35
+ ```
36
+ @inproceedings{liu2024etbench,
37
+ title={E.T. Bench: Towards Open-Ended Event-Level Video-Language Understanding},
38
+ author={Liu, Ye and Ma, Zongyang and Qi, Zhongang and Wu, Yang and Chen, Chang Wen and Shan, Ying},
39
+ booktitle={Neural Information Processing Systems (NeurIPS)},
40
+ year={2024}
41
+ }
42
+ ```