My research lies at the intersection of computer vision and deep learning, with a focus on developing robust and controllable generative models for diverse applications in image editing, restoration, and composition.
I presented our research prototype on relighting and harmonization at Adobe MAX Sneak 2024. Link: #ProjectPerfectBlend .
04-2024
I received the Pearl Brownstein Doctoral Research Award from NYU CSE for “doctoral research which shows the greatest promise”.
02-2024
Our work on lighting-aware background replacement has been accepted to CVPR2024.
12-2023
Passed my Ph.D thesis defense :)
09-2023
Our work on keypoint augmented self-supervised learning has been accepted to NeurIPS2023.
07-2023
Our work on structure guided diffusion model for deblurring has been accepted to ICCV2023.
07-2023
Our work on data synthesis for microscopy segmentation has been accepted to MICCAI DALI.
05-2023
Starting my internship at Adobe.
10-2022
I received a Scholar Award from NeurIPS2022.
09-2022
Our work on spatiotemporal representation learning has been accepted to NeurIPS2022 (oral).
08-2022
I gave a talk on my PhD research on image-to-image translation at Luma seminar, Google Research.
07-2022
I gave an invited presentation on longitudinal neuroimage analysis at Stanford Research Institute & Computational Neuroimage Science Laboratory. Milestone: my first in-person talk :p
06-2022
Starting my internship at Computational Imaging (LUMA) Team, Google Research.
04-2022
Guest lecture on "Deep Learning for Computer Vision" for NYU Tandon CS-GY 6643 Computer Vision.
07-2021
Our work on spatiotemporal brain atlas synthesis has been accepted to ICCV2021.
06-2021
Our work on diffusion-weighted brain image synthesis has been accepted to MICCAI2021 (oral).
We introduce a high-fidelity portrait shadow removal model that can effectively enhance the image of a portrait by predicting its appearance under disturbing shadows and highlights.
We introduce Relightful Harmonization, a lighting-aware diffusion model designed to seamlessly harmonize sophisticated lighting effect for the foreground portrait using any background image.
Image-conditioned Diffusion Probablistic Models (icDPMs) for restoration work well on benchmarks but not real images. We introduce a simple yet effective structure guidance that leads to significantly better visual quality on unseen images.
CNN-extracted global and local features are limited in capturing long-range spatial dependencies that are essential in biological anatomy. We present a keypoint-augmented fusion layer that extracts representations preserving both short- and long-range self-attention in a self-supervised manner.
We propose a local and multi-scale spatiotemporal representation learning method for image-to-image architectures trained on longitudinal (non i.i.d) images.
Based on an underlying assumption that morphological shape is consistent across imaging sites,
we propose a segmentation-renormalized image translation framework to reduce inter-scanner heterogeneity while preserving anatomical layout.
We propose a generative adversarial translation framework for high-quality DW image estimation with arbitrary Q-space sampling
given commonly acquired structural images.
We reformulate deformable registration and conditional
template estimation as an adversarial game wherein we encourage realism in the
moved templates with a generative adversarial registration framework
conditioned on flexible image covariates.