Roni Sengupta · SPIN Lab & UNC Chapel Hill
Roni Sengupta

Roni Sengupta

Assistant Professor · Department of Computer Science · UNC Chapel Hill

Spatial & Physical Intelligence (SPIN) Lab

I lead the SPIN Lab at UNC Chapel Hill. My research lies at the intersection of Computer Vision and Computer Graphics, mainly centered around 3D Vision and Computational Photography. My lab is particularly interested in developing AI techniques for understanding spatial and physical properties from images and videos - geometry, motion, material reflectance, material deformation properties, and lighting (Inverse Physics). We solve Inverse Physics to advance applications in Immersive Media (AR/VR, telepresence, content creation), Healthcare (medical image computing, surgical robotics), Robotics and Physical Sciences (material design and engineering).

Prior to UNC, I was a Postdoctoral Research Associate at University of Washington, working with Prof. Steve Seitz, Prof. Brian Curless, and Prof. Ira Kemelmacher-Shlizerman in the UW Reality Lab / GRAIL (2019–2022). I completed my Ph.D. at University of Maryland – College Park (2013–2019), advised by Prof. David Jacobs, and my undergraduate degree in Electronics & Tele-Communication Engineering from Jadavpur University, Kolkata, India (2009–2013). I have also spent time working with researchers at NVIDIA Research, Snapchat Research, The Weizmann Institute of Science, and TU Dortmund.

📧 ronisen at cs.unc.edu 🏛 Sitterson Hall 255, UNC Chapel Hill Google Scholar @SenguptRoni

Awards & Honors

NIH NIBIB Trailblazer Award for New and Early Stage Investigators — 2024
UNC Junior Faculty Development Award — 2024
UNC CS Student Association Excellence in Teaching Award — 2023
CVPR Best Student Paper Honorable Mentions — 2021

Research

Research Overview

Our research sits at the intersection of Computer Vision and Computer Graphics, with a focus on 3D Vision and Computational Photography. We develop AI techniques that enable machines to understand the spatial and physical structure of the world from visual data.

Our work spans two complementary directions: explicit estimation of spatial/physical properties through inverse problems leveraging foundation AI models, and implicit manipulation of these properties using generative AI. We advance both fundamental methods and practical applications across VFX, AR/VR, robotics, telepresence, and healthcare.

Currently our lab's efforts center around four synergistic research themes:

Inverse Rendering

Inverse Rendering

Inverse rendering seeks to recover the physical properties of a scene — geometry, material reflectance, and lighting — from images or videos. We develop explicit estimation methods using neural networks and implicit understanding frameworks that enable intuitive manipulation of disentangled scene components. We leverage foundation models and generative priors to build robust, generalizable techniques.

Relevant Publications

Endoscopy

3D Perception from Endoscopy

3D perception in endoscopy unlocks critical applications in medical imaging: automated measurement of organ geometry, enhanced visualization for diagnosis, and guidance for robotic surgery. The task is extremely challenging due to complex lighting effects — near-field illumination, global light transport, specular highlights, and subsurface scattering. We focus on monocular depth estimation and SLAM methods by explicitly modeling light propagation, in close collaboration with experts in medical imaging, robotics, gastroenterology, and otolaryngology.

Relevant Publications

Inverse Physics

Inverse Physics

Inverse physics involves recovering an object's 3D geometry and physical properties — such as material stiffness or initial force conditions — from sparse or single-view videos. This task is highly ill-posed and requires careful optimization. Our research addresses these challenges by designing effective optimization techniques and learning-based priors that improve inverse estimation under limited observations.

Relevant Publications

Generative Facial Editing

Generative Editing

We explore generative model-based approaches to high-quality facial attribute editing, including aging/de-aging, relighting, harmonization, and identity-preserving modifications for visual effects and creative applications. Our focus is on developing personalized, training-free methods that resolve the long-standing trade-off between inversion accuracy and editability in generative image editing frameworks.

Relevant Publications

We are grateful for the generous support from our sponsors.

NIH NIBIB Lenovo NIH NIMHD NCDOT NC Innovation

Publications

arXiv Pre-prints
Over++: Generative Video Compositing for Layer Interaction Effects
Luchao Qi, Jiaye Wu, Jun Myeong Choi, Cary Philips, Roni Sengupta, Dan B Goldman
arXiv 2025
Generates physically plausible environmental effects between foreground objects and background scenes, supporting prompt- and mask-based control.
GAINS: Gaussian-based Inverse Rendering from Sparse Multi-View Captures
Patrick Noras, Jun Myeong Choi, Didier Stricker, Pieter Peers, Roni Sengupta
arXiv 2025
Inverse rendering of specular objects from sparse views using learning-based priors.
2026
ProJo4D: Progressive Joint Optimization for Sparse-View Inverse Physics Estimation
Daniel Rho, Jun Myeong Choi, Biswadip Dey, Roni Sengupta
TMLR 2026
Recovers 3D shape and physical behavior of deformable objects from sparse-view inputs using a progressive joint-optimization framework for inverse physics estimation.
HarmoVid: Relightful Video Portrait Harmonization
Jun Myeong Choi, Jae Shin Yoon, Luchao Qi, Roni Sengupta, Joon-Young Lee
CVPR 2026
Repurposing I2V Diffusion models for harmonizing the lighting of a foreground video to match a target background scene, adjusting shadows, color tone, and illumination intensity.
Prune-Then-Plan: Step-Level Calibration for Stable Frontier Exploration in Embodied Question Answering
Noah Frahm, Prakrut Patel, Yue Zhang, Shoubin Yu, Mohit Bansal, Roni Sengupta
CVPR 2026 Findings
Uses VLM to prune bad next-steps/frontier/waypoints and then chooses the nearest frontier to maximize coverage in embodied Q&A tasks.
GLOW: Global Illumination-Aware Inverse Rendering of Indoor Scenes Captured with Dynamic Co-Located Light Camera
Jiaye Wu, Saeed Hadadan, Geng Lin, Matthias Zwicker, David Jacobs, Roni Sengupta
CVPR 2026 Findings
Global and near-field illumination-aware neural inverse rendering for recovering geometry, albedo, and roughness from co-located light+camera captures.
TalkingHeadBench: A Multi-Modal Benchmark & Analysis of Talking-Head DeepFake Detection
Xinqi (Ana) Xiong*, Prakrut Patel*, Qingyuan Fan*, Amisha Wadhwa*, Sarathy Selvam, Xiao Guo, Luchao Qi, Xiaoming Liu, Roni Sengupta
WACV 2026
2025
NFL-BA: Improving Endoscopic SLAM with Near-Field Light Bundle Adjustment
Andrea Dunn Beltran*, Daniel Rho*, Marc Niethammer, Roni Sengupta
NeurIPS 2025
Introduces a novel Bundle Adjustment loss using lighting cues for improving pose and map estimation in dense visual SLAM for endoscopy.
The Aging Multiverse: Generating Condition-Aware Facial Aging Tree via Training-Free Diffusion
Bang Gong*, Luchao Qi*, Jiaye Wu, Zhicheng Fu, Chunbo Song, David Jacobs, John Nicholson, Roni Sengupta
SIGGRAPH Asia 2025
Uses training-free diffusion to generate diverse, realistic aging paths from a single face, conditioned on health, lifestyle, and environment.
MyTimeMachine: Personalized Facial Age Transformation
Luchao Qi, Jiaye Wu, Bang Gong, Annie N. Wang, David Jacobs, Roni Sengupta
SIGGRAPH 2025 (Journal Track) · ACM ToG
Personalizes a pre-trained global aging prior using 50 personal selfies, enabling high-fidelity age regression and progression with identity preservation.
ScribbleLight: Single Image Indoor Relighting with Scribbles
Jun Myeong Choi, Annie N. Wang, Pieter Peers, Anand Bhattad, Roni Sengupta
CVPR 2025
A generative model that supports local fine-grained control of lighting effects through scribbles describing changes in lighting.
My3DGen: A Scalable Personalized 3D Generative Model
Luchao Qi, Jiaye Wu, Annie Wang, Shengze Wang, Roni Sengupta
WACV 2025 · Oral
Parameter-efficient approach for personalized 3D generative priors: only 0.6M parameters vs. full fine-tuning of 31M.
Continual Learning of Personalized Generative Face Models with Experience Replay
Annie Wang, Luchao Qi, Roni Sengupta
WACV 2025
A continual learning approach for updating personalized 2D and 3D generative face models without forgetting past representations.
2024
Leveraging Near-Field Lighting for Monocular Depth Estimation from Endoscopy Videos
Akshay Paruchuri, Samuel Ehrenstein, Shuxian Wang, Inbar Fried, Stephen M. Pizer, Marc Niethammer, Roni Sengupta
ECCV 2024
Models near-field lighting as Per-Pixel Shading (PPS) features for depth refinement on clinical endoscopy videos with self-supervision.
Personalized Video Relighting With an At-Home Light Stage
Jun Myeong Choi, Max Christman, Roni Sengupta
ECCV 2024
Builds HQ face relighting model by recording a person watching YouTube videos on their monitor instead of expensive data capture.
NePhi: Neural Deformation Fields for Approximately Diffeomorphic Medical Image Registration
Lin Tian, Hastings Greer, Raúl San José Estépar, Roni Sengupta, Marc Niethammer
ECCV 2024
Neural deformation field for medical image registration with reduced memory, enabling high-resolution registration.
Structure-preserving Image Translation for Depth Estimation in Colonoscopy
Shuxian Wang, Akshay Paruchuri, Zhaoxi Zhang, Sarah K McGill, Roni Sengupta
MICCAI 2024 · Oral
GAN + structure-preserving loss for sim2real transfer producing SOTA depths on clinical colonoscopy data.
Building Secure and Engaging Video Communication by Using Monitor Illumination
Jun Myeong Choi, Johnathan Leung, Noah Frahm, Max Christman, Gedas Bertasius, Roni Sengupta
CVPR 2024 Workshop on Multimedia Forensics
Uses light reflected from the monitor to detect if a person in a video call is real/live or deepfake.
Bringing Telepresence to Every Desk
Shengze Wang, Ziheng Wang, Ryan Schmelzle, Liujie Zheng, Youngjoong Kwon, Roni Sengupta, Henry Fuchs
IEEE TVCG 2024
Novel system rendering high-quality novel views from 4 RGBD cameras in a tele-conferencing setup with a multiview point cloud rendering algorithm.
Universal Guidance for Diffusion Models
Arpit Bansal, Hong-Min Chu, Avi Schwarzschild, Roni Sengupta, Micah Goldblum, Jonas Geiping, Tom Goldstein
ICLR 2024
Enables controlling diffusion models by arbitrary guidance modalities without retraining use-specific components.
Motion Matters: Neural Motion Transfer for Better Camera Physiological Sensing
Akshay Paruchuri, Xin Liu, Yulu Pan, Shwetak Patel, Daniel McDuff*, Roni Sengupta*
WACV 2024 · Oral (2.5% acceptance rate)
Neural Motion Transfer as data augmentation for PPG signal estimation from facial videos, improving up to 75% over SOTA on five benchmark datasets.
Joint Depth Prediction and Semantic Segmentation with Multi-View SAM
Mykhailo Shvets, Dongxu Zhao, Marc Niethammer, Roni Sengupta, Alexander C. Berg
WACV 2024
Generalized SAM features build a richer cost volume for MVS; predicted depth serves as a rich prompt for semantic segmentation.
2023
rPPG-Toolbox: Deep Remote PPG Toolbox
Xin Liu, Akshay Paruchuri, Girish Narayanswamy, Xiaoyu Zhang, Jiankai Tang, Yuzhe Zhang, Yunato Wang, Roni Sengupta, Shwetak Patel, Daniel McDuff
NeurIPS 2023 · Datasets & Benchmarks Track
Comprehensive toolbox for rPPG models with support for public benchmark datasets, data augmentation, and systematic evaluation.
MVPSNet: Fast Generalizable Multi-view Photometric Stereo
Dongxu Zhao, Daniel Lichy, Pierre-Nicolas Perrin, Jan-Michael Frahm, Roni Sengupta
ICCV 2023
Generalized multi-view photometric stereo; same reconstruction quality while being 400x faster than per-scene optimization.
Measured Albedo in the Wild: Filling the Gap in Intrinsics Evaluation
Jiaye Wu, Sanjoy Chowdhury, Hariharmano Shanmugaraja, David Jacobs, Roni Sengupta
ICCP 2023
New dataset (MAW) and three new metrics complementing WHDR for comprehensive albedo evaluation of inverse rendering algorithms.
A Surface-normal Based Neural Framework for Colonoscopy Reconstruction
Shuxian Wang, Yubo Zhang, Sarah K McGill, Julian G Rosenman, Jan-Michael Frahm, Roni Sengupta, Stephen M Pizer
IPMI 2023
SLAM + near-field Photometric Stereo for 3D colon reconstruction from colonoscopy videos.
2022
Towards Unified Keyframe Propagation Models
Patrick Esser, Peter Michael, Roni Sengupta
CVPRW 2022 · AI for Content Creation
Two-stream approach for video in-painting: high-frequency features interact locally and low-frequency features interact globally via attention.
Real-Time Light-Weight Near-Field Photometric Stereo
Daniel Lichy, Roni Sengupta, David Jacobs
CVPR 2022
Significantly faster and memory-efficient near-field photometric stereo with better quality than SOTA methods.
Robust High-Resolution Video Matting with Temporal Guidance
Peter Lin, Linjie Yang, Imran Saleemi, Roni Sengupta
WACV 2022
Alpha matting on videos via recurrent architecture. No background image or manual annotation required.
2021
A Light Stage on Every Desk
Roni Sengupta, Brian Curless, Ira Kemelmacher-Shlizerman, Steve Seitz
ICCV 2021
Learns a personalized relighting model by capturing a person watching YouTube videos, enabling relighting during video calls.
Shape and Material Capture at Home
Daniel Lichy, Jiaye Wu, Roni Sengupta, David Jacobs
CVPR 2021
High-quality photometric stereo with a simple flashlight via progressive multi-scale refinement of geometry and reflectance.
Real-Time High Resolution Background Matting
Peter Lin*, Andrey Ryabtsev*, Roni Sengupta, Brian Curless, Steve Seitz, Ira Kemelmacher-Shlizerman
CVPR 2021 · Oral · Best Paper Candidate · Best Student Paper Honorable Mentions
Background replacement at 30fps on 4K and 60fps on HD. Alpha matte extracted at low-res and selectively refined with patches.
2020
Lifespan Age Transformation Synthesis
Roy Or-El, Roni Sengupta, Ohad Fried, Eli Shechtman, Ira Kemelmacher-Shlizerman
ECCV 2020
Age transformation from 0–70 modeled by 10 anchor age classes with interpolation in the latent space.
Background Matting: The World is Your Green Screen
Roni Sengupta, Vivek Jayaram, Brian Curless, Steve Seitz, Ira Kemelmacher-Shlizerman
CVPR 2020
Extracts alpha matte without trimap annotation by leveraging a captured background image.
2019
Neural Inverse Rendering of an Indoor Scene from a Single Image
Roni Sengupta, Jinwei Gu, Kihwan Kim, Guilin Liu, David Jacobs, Jan Kautz
ICCV 2019
Self-supervision via a Residual Appearance Renderer that casts shadows, inter-reflections, and near-field lighting given predicted normals and albedo.
Pre-2019
SfSNet: Learning Shape, Reflectance and Illuminance of Faces in the Wild
Roni Sengupta, Angjoo Kanazawa, Carlos D. Castillo, David Jacobs
CVPR 2018 · Spotlight
Decomposes unconstrained faces into surface normal, albedo, and spherical harmonics lighting. Extended to TPAMI 2020 with SfSMesh for 3D face reconstruction.
Solving Uncalibrated Photometric Stereo Using Fewer Images by Jointly Optimizing Low-rank Matrix Completion and Integrability
Roni Sengupta, Hao Zhou, Walter Forkel, Ronen Basri, Tom Goldstein, David Jacobs
JMIV 2018
Uncalibrated Photometric Stereo with as few as 4–6 images via rank-constrained nonlinear optimization with ADMM.
A New Rank Constraint on Multi-view Fundamental Matrices, and its Application to Camera Location Recovery
Roni Sengupta, Tal Amir, Meirav Galun, Amit Singer, Tom Goldstein, David Jacobs, Ronen Basri
CVPR 2017 · Spotlight
Proves stacked fundamental matrices have rank 6; introduces ADMM-based optimization using this constraint to improve Structure-from-Motion.
Frontal to Profile Face Verification in the Wild
Roni Sengupta, Jun-Cheng Chen, Carlos D. Castillo, Vishal M. Patel, Rama Chellappa, David Jacobs
WACV 2016
Introduces the CFP dataset for frontal vs. profile face verification in the wild; widely used for face verification and pose-synthesis research.
No image
A Frequency Domain Approach to Silhouette Based Gait Recognition
Roni Sengupta, Udit Halder, Rameshwar Panda, Ananda S Chowdhury
NCVPRIPG 2013

Team

PhD Students

Luchao Qi
4th year
Jun Myeong Choi
4th year
Daniel Rho
2nd year
Ana Xiong
2nd year
Noah Frahm
2nd year

MS Students

Prakrut Patel
Prakrut Patel
Aarav Mehta
Aarav Mehta
Katie Sugg
Katie Sugg

Undergraduate Students

Shikha Vyas
Shikha Vyas
Shreya Ravi
Shreya Ravi
Ryan Krasinski
Ryan Krasinski
Alan Liu
Alan Liu
Haoze Li
Haoze Li
Matthew Thorton
Matthew Thorton

Alumni

Former PhD & MS Students

👤
Jiaye Wu
PhD · Now at Amazon
(w/ David Jacobs, UMD)
Daniel Lichy
PhD · Now at Kitware
(w/ David Jacobs, UMD)
Dongxu Zhao
Dongxu Zhao
PhD · Now at Google
(w/ Jan-Michael Frahm)
Patrick Noras
Patrick Noras
MS · DAAD Fellow
Annie Wang
Annie Wang
MS · Now at Databricks
Bang Gong
Bang Gong
MS

Former Undergraduates

Teaching

Lab Photos