Learning Latent Proxies for Controllable Single-Image Relighting
Abstract
Single-image relighting is highly under-constrained: small illumination changes can produce large, nonlinear variations in shading, shadows, and specularities, while geometry and materials remain unobserved. Existing diffusion-based approaches either rely on intrinsic or G-buffer pipelines that require dense and fragile supervision, or operate purely in latent space without physical grounding, making fine-grained control of direction, intensity, and color unreliable. We observe that a full intrinsic decomposition is unnecessary and redundant for accurate relighting. Instead, sparse but physically meaningful cues, indicating where illumination should change and how materials should respond, are sufficient to guide a diffusion model. Based on this insight, we introduce LightCtrl that integrates physical priors at two levels: a few-shot latent proxy encoder that extracts compact material-geometry cues from limited PBR supervision, and a lighting-aware mask that identifies sensitive illumination regions and steers the denoiser toward shading relevant pixels. To compensate for scarce PBR data, we refine the proxy branch using a DPO-based objective that enforces physical consistency in the predicted cues. We also present ScaLight, a large-scale object-level dataset with systematically varied illumination and complete camera-light metadata, enabling physically consistent and controllable training. Across object and scene level benchmarks, our method achieves photometrically faithful relighting with accurate continuous control, surpassing prior diffusion and intrinsic-based baselines, including gains of up to +2.4 dB PSNR and 35% lower RMSE under controlled lighting shifts.
Community
We introduce LightCtrl, a diffusion based relighting framework that replaces dense intrinsic decomposition with sparse physical priors, which have a few-shot latent proxy encoder for material-geometry cues and a lighting-aware mask for spatially guided denoising. We also release ScaLight, a large scale object-level relighting dataset with full camera-light metadata.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Stronger Semantic Encoders Can Harm Relighting Performance: Probing Visual Priors via Augmented Latent Intrinsics (2026)
- SyncLight: Controllable and Consistent Multi-View Relighting (2026)
- Light Up Your Face: A Physically Consistent Dataset and Diffusion Model for Face Fill-Light Enhancement (2026)
- PhaSR: Generalized Image Shadow Removal with Physically Aligned Priors (2026)
- Ctrl&Shift: High-Quality Geometry-Aware Object Manipulation in Visual Generation (2026)
- PI-Light: Physics-Inspired Diffusion for Full-Image Relighting (2026)
- LuxRemix: Lighting Decomposition and Remixing for Indoor Scenes (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper