
VideoVerse: How Far is Your T2V Generator from a World Model?
Abstract
The recent rapid advancement of Text-to-Video (T2V) generation technologies, which are critical to build "world models", makes the existing benchmarks increasingly insufficient to evaluate state-of-the-art T2V models. First, current evaluation dimensions, such as per-frame aesthetic quality and temporal consistency, are no longer able to differentiate state-of-the-art T2V models. Second, event-level temporal causality, which not only distinguishes video from other modalities but also constitutes a crucial component of world models, is severely underexplored in existing benchmarks. Third, existing benchmarks lack a systematic assessment of world knowledge, which are essential capabilities for building world models. To address these issues, we introduce VideoVerse, a comprehensive benchmark that focuses on evaluating whether a T2V model could understand complex temporal causality and world knowledge in the real world. We collect representative videos across diverse domains (e.g., natural landscapes, sports, indoor scenes, science fiction, chemical and physical experiments) and extract their event-level descriptions with inherent temporal causality, which are then rewritten into text-to-video prompts by independent annotators. For each prompt, we design a suite of binary evaluation questions from the perspective of dynamic and static properties, with a total of ten carefully defined evaluation dimensions. In total, our VideoVerse comprises 300 carefully curated prompts, involving 815 events and 825 binary evaluation questions. Consequently, a human preference aligned QA-based evaluation pipeline is developed by using modern vision-language models. Finally, we perform a systematic evaluation of state-of-the-art open-source and closed-source T2V models on VideoVerse, providing in-depth analysis on how far the current T2V generators are from world models.
Evolution Demos
VideoVerse Benchmark Leaderboard (Evaluated By Gemini2.5 Pro)
Model | Overall | Dynamic | Static | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Event Following |
Camera Control |
Interaction | Mechanics | Material Properties |
Natural Constra. |
Common Sense |
Attr. Correct. |
2D Layout |
3D Depth |
||
Open-Source Models | |||||||||||
CogVideoX1.5 (S) | 922 | 424 | 37 | 37 | 25 | 26 | 36 | 41 | 178 | 66 | 52 |
CogVideoX1.5 (L) | 916 | 426 | 38 | 38 | 28 | 22 | 38 | 38 | 183 | 58 | 47 |
SkyReels-V2 (S) | 963 | 484 | 43 | 37 | 30 | 22 | 32 | 43 | 161 | 61 | 50 |
SkyReels-V2 (L) | 997 | 511 | 37 | 42 | 33 | 24 | 36 | 36 | 169 | 62 | 47 |
Wan2.1-14B | 998 | 496 | 43 | 34 | 32 | 24 | 35 | 46 | 168 | 68 | 52 |
Hunyuan | 923 | 446 | 39 | 32 | 34 | 25 | 37 | 42 | 160 | 60 | 48 |
OpenSora2.0 | 1015 | 482 | 48 | 36 | 29 | 27 | 48 | 50 | 182 | 62 | 51 |
Wan2.2-A14B | 1112 | 567 | 61 | 36 | 39 | 30 | 37 | 44 | 185 | 64 | 49 |
Closed-Source Models | |||||||||||
Minimax-Hailuo | 1241 | 623 | 76 | 44 | 42 | 36 | 55 | 53 | 188 | 69 | 55 |
Veo-3 | 1334 | 680 | 77 | 54 | 50 | 36 | 68 | 58 | 188 | 68 | 55 |
BibTeX
@article{YourPaperKey2024,
title={Your Paper Title Here},
author={First Author and Second Author and Third Author},
journal={Conference/Journal Name},
year={2024},
url={https://your-domain.com/your-project-page}
}