Scene |
Depth Map |
Direct Component |
Global Component |
|
Most active scene recovery techniques assume that a scene point is
illuminated only directly by the illumination source. Consequently,
global illumination effects due to inter-reflections, sub-surface
scattering and volumetric scattering introduce strong biases in the
recovered scene shape. Our goal is to recover scene properties in
the presence of global illumination. To this end, we study the
interplay between global illumination and the depth cue of illumination
defocus. By expressing both these effects as low pass filters, we derive
an approximate invariant that can be used to separate them without
explicitly modeling the light transport. This is directly useful in any
scenario where limited depth-of-field devices (such as projectors)
are used to illuminate scenes with global light transport and significant
depth variations. We show two applications: (a) accurate depth recovery
in the presence of global illumination, and (b) factoring out the effects
of defocus for correct direct-global separation in large depth scenes.
We demonstrate our approach using scenes with complex shapes, reflectances,
textures and translucencies.
|
Publications
"A Combined Theory of Defocused Illumination and Global Light Transport"
Mohit Gupta, Yuandong Tian, Srinivasa G. Narasimhan, Li Zhang,
International Journal of Computer Vision (IJCV),
June 2012.
[PDF]
"(De) Focusing on Global Light Transport for Active Scene Recovery"
Mohit Gupta, Yuandong Tian, Srinivasa G. Narasimhan, Li Zhang,
IEEE Computer Vision and Pattern Recognition (CVPR),
2009.
[PDF]
"Relationship between projector defocus and global
illumination for statistically-modeled scenes."
Yuandong Tian, Mohit Gupta, Srinivasa G. Narasimhan, Li Zhang,
Technical Report CMU-RI-TR-09-10, Carnegie Mellon University,
March 2009.
|
Talk and Video
(Video Result Playlist)
|
Results
|
The 'Atlas and Candle' scene.
|
|
The 'Organic' scene.
|
|
Scene with complex BRDFs and geometry.
|
|
The 'Candles and Soaps' scene.
|
|
The Real vs. Fake' scene.
|
|
Acknowledgements
This research was supported in parts by
an ONR grant N00014-08-1-0330 and NSF awards CCF-0541307,
IIS-0643628 and IIS-0845916.
|
|