Lighting-grounded Video Generation with Renderer-based Agent Reasoning
Abstract
LiVER presents a diffusion-based framework for scene-controllable video generation that disentangles 3D scene properties through explicit conditioning and automated user instruction translation.
Diffusion models have achieved remarkable progress in video generation, but their controllability remains a major limitation. Key scene factors such as layout, lighting, and camera trajectory are often entangled or only weakly modeled, restricting their applicability in domains like filmmaking and virtual production where explicit scene control is essential. We present LiVER, a diffusion-based framework for scene-controllable video generation. To achieve this, we introduce a novel framework that conditions video synthesis on explicit 3D scene properties, supported by a new large-scale dataset with dense annotations of object layout, lighting, and camera parameters. Our method disentangles these properties by rendering control signals from a unified 3D representation. We propose a lightweight conditioning module and a progressive training strategy to integrate these signals into a foundational video diffusion model, ensuring stable convergence and high fidelity. Our framework enables a wide range of applications, including image-to-video and video-to-video synthesis where the underlying 3D scene is fully editable. To further enhance usability, we develop a scene agent that automatically translates high-level user instructions into the required 3D control signals. Experiments show that LiVER achieves state-of-the-art photorealism and temporal consistency while enabling precise, disentangled control over scene factors, setting a new standard for controllable video generation.
Community
Accepted to CVPR 2026
TL;DR: LiVER is a controllable video diffusion framework that generates videos from explicit 3D scene conditions such as layout, lighting, and camera motion. It combines a densely annotated dataset, a lightweight conditioning module, and a scene agent that converts user instructions into editable 3D controls, achieving strong realism and temporal consistency with much more precise scene control.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Motion Forcing: A Decoupled Framework for Robust Video Generation in Motion Dynamics (2026)
- Tri-Prompting: Video Diffusion with Unified Control over Scene, Subject, and Motion (2026)
- CamLit: Unified Video Diffusion with Explicit Camera and Lighting Control (2026)
- NavCrafter: Exploring 3D Scenes from a Single Image (2026)
- GO-Renderer: Generative Object Rendering with 3D-aware Controllable Video Diffusion Models (2026)
- DCDM: Divide-and-Conquer Diffusion Models for Consistency-Preserving Video Generation (2026)
- HorizonForge: Driving Scene Editing with Any Trajectories and Any Vehicles (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Get this paper in your agent:
hf papers read 2604.07966 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper