Stable Cinemetrics: Structured Taxonomy and Evaluation for Professional Video Generation¶
Conference: NeurIPS 2025 arXiv: 2509.26555 Code: Project Page Area: Video Generation / Evaluation Benchmark Keywords: video generation evaluation, cinematic taxonomy, professional video control, human evaluation, VLM evaluator
TL;DR¶
This paper proposes SCINE (Stable Cinemetrics), the first structured evaluation framework targeting professional video production. It defines a hierarchical taxonomy with 76 fine-grained cinematic control nodes, accompanied by large-scale professional annotation (80+ film practitioners, 20K+ videos, 248K annotations), revealing significant deficiencies of current state-of-the-art T2V models in professional cinematic control.
Background & Motivation¶
Video generation models have advanced rapidly, yet existing benchmarks (e.g., VBench, VideoPhy) fail to capture the requirements of professional video production. The core gap between professional filmmaking and casual generation lies in cinematic control: professional directors require precise control over shot composition, lighting quality, action timing, and every other cinematic element, rather than simply accepting a model output of "an astronaut riding a horse."
Specific limitations of existing benchmarks include:
Lack of cinematographic depth: VBench prompts such as "A man is walking" omit professionally essential information including character appearance, scene setup, and camera motion.
Coarse evaluation dimensions: Most benchmarks assess only overall prompt adherence, without attributing performance to specific control parameters.
Static design: Fixed prompt sets cannot scale with improving model capabilities.
Lack of professional validation: Automatic metrics are poorly aligned with expert human judgment.
The authors' core argument is that the cinematic shot is the atomic unit of filmmaking (averaging 5–10 seconds, which aligns naturally with the temporal limits of current models). A single shot involves a large number of mutually independent control parameters, providing a natural basis for structured evaluation.
Method¶
Overall Architecture¶
SCINE consists of three components: 1. Taxonomy: A hierarchical control tree with 4 pillars and 76 leaf nodes. 2. Benchmark Prompts: Two prompt types (narrative scripts + visual elaborations) simulating professional workflows. 3. Evaluation Pipeline: Automatic taxonomy mapping → question generation → large-scale human/automatic evaluation.
Key Designs¶
-
Four Taxonomic Pillars (76 control nodes):
- Setup: Scene texture, geometry, set design, props, background, character appearance, etc.—"everything visible in the frame."
- Camera: Intrinsics (focal length, depth of field, ISO), extrinsics (angle, height), trajectory (camera movement, tracking), and creative intent (composition, shot size).
- Lighting: Light source type, color temperature, lighting conditions, effects, position, and advanced controls.
- Events: Action types (standalone/interactive), emotions (explicit/implicit), dialogue, temporal presentation (atomic/causal/concurrent/cyclical), rhythm, and narrative structure.
Design principle: A hierarchical tree structure ensures inter-branch independence (adjusting depth of field does not affect camera movement), supports multi-level abstraction, and facilitates extensibility.
-
Prompt Design Pipeline:
- SCINE-Scripts: Seed prompts are created in collaboration with professional screenwriters; Events taxonomy nodes are sampled and narrative scripts are generated via LLM. t-SNE analysis confirms high distributional overlap with real screenplays.
- SCINE-Visuals: Control nodes sampled from Camera, Lighting, and Setup taxonomy branches are injected into Scripts, enabling structured prompt augmentation (rather than unconstrained LLM expansion).
- A single script can yield multiple visual interpretations (e.g., the same "man serving dinner to family" scene can be paired with shallow depth of field + warm lighting vs. deep depth of field + cold lighting).
-
Automatic Taxonomy Mapping and Question Generation: Each prompt is automatically mapped to taxonomy nodes, and independent evaluation questions are generated per node. For example, a prompt mentioning "tight close-up" and "flickering" generates separate evaluation questions for Shot Size and Lighting Motion, enabling decoupled assessment of individual control nodes.
Loss & Training¶
VLM Evaluator Training: Based on Qwen-2.5-VL-7B, fine-tuned with a Bradley-Terry preference objective: - Training set: 44,062 samples; validation set: 12,763 samples. - Input: single video + prompt + evaluation question → Output: scalar score. - A linear projection head on the last-layer token produces the scalar value. - Sampled at 2 fps at native resolution; trained for 1 epoch.
Key Experimental Results¶
Main Results¶
SCINE Visuals Comparison Across Four Pillars (13 Models)
| Pillar | Top Model | Score Trend | Key Finding |
|---|---|---|---|
| Setup | WAN-14B highest | Highest absolute scores | Best-performing dimension across all models |
| Lighting | Consistent across models | Smallest spread | Natural light > artificial light |
| Camera | Universally low | Narrow spread | Similar bottleneck across all models |
| Events | Largest gap | Only top-3 reliable | Most challenging dimension |
Fine-Grained Analysis of Events
| Sub-category | Performance | Notes |
|---|---|---|
| Standalone Actions | Better | Standalone > interactive actions |
| Implicit Emotions | Better | Implicit > explicit emotions |
| Atomic Events | Better | Best performance among event types |
| Dialogues | Worse | Minimax leads but gap remains large |
| Causal/Overlapping | Worse | Events requiring temporal reasoning are broadly difficult |
| Advanced Controls | Worse | Rhythm and narrative structure are the hardest to control |
Ablation Study¶
| Configuration | Key Observation | Notes |
|---|---|---|
| Basic vs. Advanced prompts | All models decline on Advanced | Largest drop in Lighting Source |
| Director prompts (joint control) | Largest drop in Camera | Multi-dimensional joint specification causes overall degradation |
| VLM scale 7B/32B/72B | No significant improvement | Zero-shot VLM alignment is poor |
| Fine-tuned 7B VLM | 72.36% accuracy | ~20% improvement over zero-shot 72B |
Key Findings¶
- Three-tier ranking: Minimax and WAN-14B lead → Luma Ray 2 / Hunyuan / WAN-1B in the middle → remaining models form a third tier.
- No model excels uniformly: Even the strongest models perform poorly on Events and Camera.
- Dutch angle is a challenge for all models; Medium-Wide and Extreme Close-up are the hardest shot sizes.
- Among lighting sources, Sunlight and Strobes perform well, while HMI and Fluorescent perform poorly.
- Causal and Sequential event performance is highly correlated (\(\rho=0.94\)), suggesting a shared temporal understanding capability.
Highlights & Insights¶
- Designing an evaluation framework from a professional filmmaking perspective bridges the significant gap between generative model assessment and real-world application.
- The 76-node taxonomy is itself a significant contribution and can serve as a reference for control dimensions in future model training.
- Taxonomy-guided prompt augmentation is more controllable and interpretable than unconstrained LLM expansion.
- The large-scale professional annotation (248K labels, 84 practitioners, ICC 80.4%) provides a solid empirical foundation for all conclusions.
- The fine-tuned VLM evaluator outperforms the zero-shot 72B model in alignment, yet the 72% accuracy indicates substantial room for improvement in automatic evaluation.
Limitations & Future Work¶
- The taxonomy is constrained by the collaborating expert network and may not cover cinematic traditions across diverse global cultures.
- Certain nodes (e.g., color temperature 2000K, ISO 800) are overly granular and difficult to accurately perceive even for human annotators.
- LLM-generated prompts may introduce biases.
- Only T2V models are evaluated; I2V models and multi-shot temporal coherence are not addressed.
- No closed-loop connection to model training is established—how to leverage these fine-grained evaluation results to guide model improvement remains unexplored.
Related Work & Insights¶
- Complementary to VBench: VBench focuses on general video quality dimensions, while SCINE targets professional cinematic control.
- MovieNet provides film-level annotations but is not designed for generative model evaluation.
- This work may inspire the incorporation of domain expertise (e.g., cinematography, photography) into evaluation framework design for other modalities.
- The structured taxonomy can also be applied to analyze cinematic diversity in video datasets or to guide video captioning.
Rating¶
- Novelty: ⭐⭐⭐⭐⭐ First systematic integration of professional filmmaking knowledge into video generation evaluation; the taxonomy design is rigorous.
- Experimental Thoroughness: ⭐⭐⭐⭐⭐ 13 models, 20K videos, 248K annotations, 84 professional annotators — impressive in both scale and quality.
- Writing Quality: ⭐⭐⭐⭐ Rich in content but lengthy; the presentation of the taxonomy could be more compact.
- Value: ⭐⭐⭐⭐⭐ Significantly advances the evaluation paradigm for video generation; both the taxonomy and benchmark are likely to be widely cited.