Skip to content

Latest commit

 

History

History
12 lines (6 loc) · 913 Bytes

File metadata and controls

12 lines (6 loc) · 913 Bytes

This folder contains the prompt sets we use to evaluate video generation performance in the paper. It includes curated subsets from public benchmarks and a challenging set we constructed with GPT-4o. Chinese translations are provided for readability.

  • VBench: 946 prompts curated from the VBench benchmark.

  • VideoGen-Eval: 400 prompts curated from VideoGen-Eval for general-purpose evaluation.

  • TA-Hard: 72 prompts generated with GPT-4o. This set emphasizes challenging cases, typically combining two subjects and incorporating uncommon actions.

For VBench and VideoGen-Eval, we rewrite the prompts to better align with our video model’s input format while preserving the original intent and difficulty.

Each prompt is available in both English and Chinese. The Chinese versions help bilingual annotators and readers.