Datasets:
🧠 EgoExoBench: A Cross-Perspective Video Understanding Benchmark
Dataset Summary
EgoExoBench is a benchmark designed to evaluate cross-perspective understanding capabilities of multimodal large models (MLLMs).
It contains synchronized and asynchronous egocentric (first-person) and exocentric (third-person) video pairs, along with multiple-choice questions that assess semantic alignment, viewpoint association, and temporal reasoning between the two perspectives.
Features
Each sample contains:
- Question: A natural-language question testing cross-perspective reasoning.
- Options: Multiple-choice answers (A/B/C/D).
- Answer: Correct answer label.
- Videos: Egocentric and Exocentric videos.
Evaluation Metric
Accuracy (%)
Data Splits
| Split | #Samples |
|---|---|
| Test | 7,330 |
Example Usage
from datasets import load_dataset
dataset = load_dataset("YourUsername/EgoExoBench")
print(dataset["test"][0])
Citation
If you use EgoExoBench in your research, please cite:
@misc{he2025egoexobench,
title={EgoExoBench: A Benchmark for First- and Third-person View Video Understanding in MLLMs},
author={Yuping He and Yifei Huang and Guo Chen and Baoqi Pei and Jilan Xu and Tong Lu and Jiangmiao Pang},
year={2025},
eprint={2507.18342},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2507.18342}
}
- Downloads last month
- 96