Real-Time Light Field Rendering for Multi-projector 3D Display

Author(s):  
Chen Qin ◽  
Chao Peng ◽  
Jia Lu ◽  
Wenhua Dou
2013 ◽  
Vol 44 (1) ◽  
pp. 667-670 ◽  
Author(s):  
Juyong Park ◽  
Dongkyung Nam ◽  
Seo Young Choi ◽  
Jin-Ho Lee ◽  
Du Sik Park ◽  
...  

2015 ◽  
Vol 11 (10) ◽  
pp. 792-799 ◽  
Author(s):  
Young Ju Jeong ◽  
Jin-Ho Lee ◽  
Yang Ho Cho ◽  
Dongkyung Nam ◽  
Du-Sik Park ◽  
...  

Author(s):  
Seok Lee ◽  
Juyong Park ◽  
Jingu Heo ◽  
Byungmin Kang ◽  
Dongwoo Kang ◽  
...  

2015 ◽  
Vol 46 (1) ◽  
pp. 346-349 ◽  
Author(s):  
Chen Su ◽  
Qing Zhong ◽  
Liang Xu ◽  
Haifeng Li ◽  
Xu Liu
Keyword(s):  

Author(s):  
T. Chlubna ◽  
T. Milet ◽  
P. Zemčík

AbstractLight field rendering is an image-based rendering method that does not use 3D models but only images of the scene as input to render new views. Light field approximation, represented as a set of images, suffers from so-called refocusing artifacts due to different depth values of the pixels in the scene. Without information about depths in the scene, proper focusing of the light field scene is limited to a single focusing distance. The correct focusing method is addressed in this work and a real-time solution is proposed for focusing of light field scenes, based on statistical analysis of the pixel values contributing to the final image. Unlike existing techniques, this method does not need precomputed or acquired depth information. Memory requirements and streaming bandwidth are reduced and real-time rendering is possible even for high resolution light field data, yielding visually satisfactory results. Experimental evaluation of the proposed method, implemented on a GPU, is presented in this paper.


2020 ◽  
Vol 6 (1) ◽  
pp. 1
Author(s):  
Enrique Canessa ◽  
Livio Tenze

We developed a method to transform stereoscopic two-dimensional (2D) images into holograms via unsupervised morphing deformations between left (L) and right (R) input images. By using robust DeepFlow and light-field rendering algorithms, we established correlations between a 2D scene and its three-dimensional (3D) display on a Looking Glass HoloPlay monitor. The possibility of applying this method, together with a lookup table for multi-view glasses-free 3D streaming with a stereo webcam, was also analyzed.


Author(s):  
Takasuke Nagai ◽  
Munekazu Date ◽  
Shinya Shimizu ◽  
Hideaki Kimata

Author(s):  
Ying Yuan ◽  
Xiaorui Wang ◽  
Yang Yang ◽  
Hang Yuan ◽  
Chao Zhang ◽  
...  

Abstract The full-chain system performance characterization is very important for the optimization design of an integral imaging three-dimensional (3D) display system. In this paper, the acquisition and display processes of 3D scene will be treated as a complete light field information transmission process. The full-chain performance characterization model of an integral imaging 3D display system is established, which uses the 3D voxel, the image depth, and the field of view of the reconstructed images as the 3D display quality evaluation indicators. Unlike most of the previous research results using the ideal integral imaging model, the proposed full-chain performance characterization model considering the diffraction effect and optical aberration of the microlens array, the sampling effect of the detector, 3D image data scaling, and the human visual system, can accurately describe the actual 3D light field transmission and convergence characteristics. The relationships between key parameters of an integral imaging 3D display system and the 3D display quality evaluation indicators are analyzed and discussed by the simulation experiment. The results will be helpful for the optimization design of a high-quality integral imaging 3D display system.


2021 ◽  
Vol 52 (1) ◽  
pp. 369-372
Author(s):  
Chen Gao ◽  
Yifan (Evan) Peng ◽  
Haifeng Li ◽  
Xu Liu

Sign in / Sign up

Export Citation Format

Share Document