Scientist+Engineer interested in understanding and maximizing the perceived quality of visual experiances. I currently work on designing 📸 Algorithms at Apple .
I did my PhD under the mentorship of Piotr Didyk, working towards realizing the dream of real-time AR/VR that is visually indistinguishable from the real world.
I have also worked with the Applied Perception Science and Display Systems Research teams at Meta on real-time perceptually optimized computational display algorithms for AR/VR.
During my Master's, I was a research fellow at KAIST, investigating how image/video enhancement neural networks understand visual quality, and how we can teach them to perceive it the same way humans do.
My primary interests lie in the intersection of human vision science and real-time computer graphics. I work on understanding,
quantifying and maximizing perceptual realism, accurate reproduction, and quality (with constituents like spatial quality, dynamic range, depth, motion and color) for real-time image/video capture (computational photography), synthesis
(rendering/graphics) and display (computational display). The long term goals I aim to push towards are:
A comprehensive understanding of how human vision understands visual realism and quality.
Real-time spectacle form-factor immersive displays (VR/AR) that are perceptually indistinguishable from the real-world.
Real-time smartphone cameras that can not only capture the world exactly as our eyes are seeing see it, but also automatically optimize for aesthetic attiributes/intent in a subjective manner.
Recent News
[Apr-2025] Our work on depth and distance perception in AR/VR to be presented at SIGGRAPH 2025.
[Oct-2024] I have joined the Camera Algorithms Team at Apple full-time.
[Mar-2024] Our work on preserving motion perception in AR/VR to be presented at SIGGRAPH 2024.
[Feb-2024] Gave an invited talk at University College London (UCL) on my work on Perceptual Optimization of Realism for real-time AR/VR. Thank you Kaan Aksit for the invitation.
[Aug-2023] Our work on ultra-fast perceptually adaptive tone mapping on VR headsets to be presented at SIGGRAPH Asia 2023.
[Oct-2022] I have joined the Vision Science team at Facebook Reality Labs (Sunnyvale, CA) as a Research Scientist Intern.
[Apr-2022] Our work on perceptual enhancement for real-time AR/VR to be presented at SIGGRAPH 2022.
Publications
Representative projects are highlighted, a few that I consider my most original and fundemental ideas.
A step towards practical and efficient real-time deep learning based image/video enhancement by leveraging the limitations of human vision. We design a perceptual model and pipeline to control the local quality and deep learning resources for image/video super resolution, such that there is no visible loss in quality, even with a very significant reduction in computational cost and runtime. The framework is general and can be applied to any deep learning based super resolution achitecture. The framework can also be applied for efficient deep denoising, frame-interpolation, and possibly generative image/video synthesis without noticeable quality loss.
ICCV 2025 - Human Vision Inspired Computer Vision Workshop
Understanding stereoscopic depth perception in Foveated AR/VR. We derive a perceptual model which demonstrates that the blur intensities applied in common foveation procedures do not affect stereoacuity. Additionally, our findings suggest that it is important to maintain high depth quality even in strongly foveated stereoscopic content.
We demonstrate that foveated rendering may inhibit motion perception, making AR/VR appear slower than it physically is. We propose the theory of Motion Metamers of human vision; videos that are structurally different from one another but indistinguishable to human peripheral vision in both spatial and motion perception. We present the first technique to synthesize motion metamers for AR/VR; all in real-time and completely unsupervised (no high-quality reference required).
An ultra-fast (under 1ms per-frame on standalone VR) framework that adaptively maintains the perceptual appearence of HDR content after tone-mapping. The framework relates human contrast perception across very different lumainances scales, and then optimizes any tone-mapping curve to minimize perceptual difference.
The fastest (200FPS at 4K) and first no-reference spatial metamers of human peripheral vision that we know of; specifically tailored for direct integration into the real-time VR foveated rendering pipeline. Save upto 40% (rendering time) over tranditional foveated rendering, without visible loss in quality.
An investigation into why the representations learned by image recognition CNNs work remarkably well as features of perceptual quality (e.g perceptual loss). We theorize that these image classification representations learn to be spectrally sensitive to the same spatial frequencies which the human visual system is most sensitive to, so they can effectively encode perceptually visible distortions.
A human contrast perception inspired spatial attention mask that makes the deep learning pipeline aware of perceptually important visual information in images.
ICCV 2019 - Learning for Computational Imaging Workshop