Paper
21 May 2004 Temporally consistent virtual camera generation from stereo image sequences
Simon R. Fox, Julien Flack, Juliang Shao, Phil Harman
Author Affiliations +
Proceedings Volume 5291, Stereoscopic Displays and Virtual Reality Systems XI; (2004) https://doi.org/10.1117/12.527895
Event: Electronic Imaging 2004, 2004, San Jose, California, United States
Abstract
The recent emergence of auto-stereoscopic 3D viewing technologies has increased demand for the creation of 3D video content. A range of glasses-free multi-viewer screens have been developed that require as many as 9 views generated for each frame of video. This presents difficulties in both view generation and transmission bandwidth. This paper examines the use of stereo video capture as a means to generate multiple scene views via disparity analysis. A machine learning approach is applied to learn relationships between disparity generated depth information and source footage, and to generate depth information in a temporally smooth manner for both left and right eye image sequences. A view morphing approach to multiple view rendering is described which provides an excellent 3D effect on a range of glasses-free displays, while providing robustness to inaccurate stereo disparity calculations.
© (2004) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Simon R. Fox, Julien Flack, Juliang Shao, and Phil Harman "Temporally consistent virtual camera generation from stereo image sequences", Proc. SPIE 5291, Stereoscopic Displays and Virtual Reality Systems XI, (21 May 2004); https://doi.org/10.1117/12.527895
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Eye

Neural networks

Cameras

Machine learning

Video

3D displays

3D image processing

RELATED CONTENT


Back to Top