999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

Light Field Virtual View Rendering Based on EPI?Representations

2016-10-13 05:54:28SUNYuleYULu
ZTE Communications 2016年3期

SUN Yule YU Lu

Abstract

Image?Based Rendering (IBR) is one powerful approach for generating virtual views. It can provide convincing animations without an explicit geometric representation. In this paper, several implementations of light field rendering are summarized from prior arts. Several characteristics, such as the regular pattern in Epipolar Plane Images (EPIs), of light field are explored with detail under 1D parallel camera arrangement. It is proved that it is quite efficient to synthesize virtual views for Super Multi?View (SMV) application, which is in the third phase of Free?Viewpoint Television (FTV). In comparison with convolutional stereo matching method, in which the intermediate view is synthesized by the two adjacent views, light field rendering makes use of more views supplied to get the high?quality views.

Keywords

depth estimate; Epipolar Plane Image (EPI); light field; view synthesis

1 Introduction

Free?Viewpoint Television (FTV) enables users to view a 3D scene by freely changing viewpoints as they do naturally in the real world. FTV is ranked as the top of visual media for its infinite number of views to display [1]. It provides a very realistic glasses?free 3D viewing without eye fatigue. Since the first phase of FTV, i.e. Multi?View Video Coding (MVC), was proposed in 2001 [2], the applications of 3DTV has developed from stereo viewing to auto?stereoscopic viewing and Super Multi?View viewing.

With the blooming of Super?Multi?View (SMV) displays, hundreds of linearly or angularly arranged, horizontal parallax ultra?dense views are required. SMV, one specific scenario in the third phase of FTV, aims at displaying hundreds of very dense and very wide baseline views [3]. Due to the limited bandwidths, it is not realistic to efficiently encode hundreds of views for viewers. In addition, different from 3D video (3DV), depth?image?based?rendering (DIBR) could not be applied for lack of depth map in SMV application.

As a simple and robust IBR method for generating new views, the 4D light field rendering simply combines and resamples the available images to generate the virtual views [4] without explicit depth map and corresponding information. Light field is a promising representation to describe the visual appearance of a scene and the special linear structure of Epipolar Plane Image (EPI) can be exploited to estimate accurate depth map which is essential for rendering quality.

The remainder of the paper is organized as follows. In section 2, we introduce the four?dimensional light field and analyze EPI in light field. Section 3 summarizes the rendering method based on light field, including depth estimation and view synthesis. Finally, conclusions are drawn in section 4.

2 FourDimensional Light Field and EPI

As we can describe each image by the location and orientation of the camera, light field is defined as the radiance to represent the related information [4]. The light field can be represented by a 5D function [f(x,y,z,θ,?)]. The five dimensional quantity function describes the flow of light at every 3D space position ([x,y,z]) for every 2D direction ([θ,?]) [5], and the intensity of the specified ray is[ f]. When there is no light dispersion in the scene, all the coincident rays along a portion of free space (between solid or refractive objects) have the same color [6]. Under these conditions, the 5D representation can be reduced to 4D function as[ L(u,v,s,t)], which is called 4D light field. 4D light field is composed of two plane ([u,v]) and ([s,t]), as shown in Fig. 1. The ([s,t]) plane contains the focal points of the views, and the ([u,v]) plane means image plane. [L(u,v,s,t)] can be viewed as an assignment of an intensity value to the ray passing through ([u,v]) and ([s,t]) [7].

It is difficult to think about 4D light field, so we fix the row value [t] to drop our visualization by one dimension. Along the [s] axis, an array of camera views is stacked. The 2D subspace [L(u,s)] is called the EPI, and an example is shown in Fig. 2. EPI consists of Epipolar lines which are the intersection of epipolar plane and the camera plane in computer vision. In the EPI, adjacent line comes from adjacent view captured by camera. Assuming the scene is Lambertian, the rays contained in such planes have same intensity and color. If cameras are linearly arranged with the same interval, the EPI seems to be composed of many straight lines with different slopes (Fig. 2). With such constraints, let us consider the light field geometry (Fig. 3). [p=(x,y,z)] is an arbitrary point in space, while A and B are the corresponding projection points on the image plane from adjacent camera. If the adjacent cameras distance is [?s], it is clear that the offset [?x=x2-x1] in the corresponding image plane can be computed using (1)

[?x=-fZ?s ] (1)

where [f]is the focal length which is the distance between two parallel planes and [Z] is the depth of the point.

Since [d=?x/?s=-f/Z] is a constant called disparity for one given 3D space point, the slope of the line which is composed of projection points from different views in EPI is related to its depth if the point is visible.

3 Light Field Rendering Analysis

Rendering methods based on light field analysis can be decomposed into 2 parts: depth estimation and view synthesis. Disparity is represented by the slope of lines in EPIs (Fig. 2). Disparity estimation is to detect the lines and calculate the corresponding slope (Sec. 3.1). Given the disparity information, virtual views are easy to synthesize by DIBR (Sec. 3.2).

3.1 Depth Estimation

The quality of the virtual views, and hence the quality of the 3DTV experience, relies heavily on the depth map accuracy [8]. Compared with traditional stereo matching methods, depth estimation through EPI characteristic can make use of more views information. It is apparent that estimating the depth from more views can be more reliable and accurate. The depth estimation methods in EPI can be classified into two kinds: global line detection and local disparity estimation. Both the two methods estimate the depth map by detecting and computing the slope in EPIs. The main difference of the two methods is whether making full use of rows in EPI.

3.1.1 Global Line Detection Method

Global line detection is the most common and most established depth estimation method in light field analysis. Similar to the stereo matching method which looks for corresponding image features in two or more camera views, the global line detection method is to determine the slope for every pixel in the EPI, i.e.to find the right corresponding projection points in different rows. The slope of the line is related to disparity described in detail in Section 2.

Although each pixel in EPI has its directionality associated with disparity, the correspondence may not be reliable if pixel is in smooth area because its surroundings are pixels with similar intensity. Based on this analysis, we give the general steps of global line detection method: 1) Extracting the characteristic pixels with high edges confidence in each EPI; 2) defining a cost function according to the coherence of intensity and gradient to determine directionality of each characteristic pixel, and 3) assigning disparity values to the pixels of the homogeneous interior region in EPI, for example, using interpolation [9] or fine to coarse procedure [10].

Compared with stereo matching, the global line detection method has higher reliability for the characteristic points. Many algorithms have been proposed to estimates disparity through locating the optimal slope of each line in EPIs [11]-[19]. Here we introduce two typical algorithms. Kim [10] proposed an algorithm which first estimates the reliable depth around object boundaries and then processes the homogeneous interior regions in a fine?to?coarse procedure, which can yield precise object contours and ensure smoothness in less detailed areas. According to the results in Fig. 4, Kims method focuses on the closeups and has better performance in homogeneous regions and areas around object contours. The latest method utilizes intensity pixel value, gradient pixel value, spatial consistency as well as reliability measure to select the best slope from a predefined set for every pixel in EPI. In this method, the spatial smoothness constraint can handle the pixels both in edge and in homogeneous regions [20].

In general, the disparity estimation is reliable by computing the slope of the line from plenty of views in the absence of occlusion. The occluded pixels at occlusion boundaries should be detected to alleviate error estimation, for example, re?projecting the EPIs and filling in occluded pixels using depth propagation [20]. However, the global line detection method is highly complex and may be challenging to run in real time with limited processing capabilities since every characteristic pixel needs to search all the rows in EPIs for choosing the optimal direction.

3.1.2 Local Disparity Estimation Method

Instead of detecting straight line with expensive matching, the local disparity estimation method by directly computing the slope in EPIs can obtain fast disparity estimation. Dansereau first proposed a directly computing method to extract depths applying gradient operators on the EPIs obtained from the light field [21]. To achieve higher quality, Wanner and Goldlucke [7] applied a structure tensor to measure pixels directions on a 3*3 stencil in EPI. The structure tensor is also used in 4D field to estimate depth [22]. The structure tensor J applied in EPI plane (x, s) is presented in (2), where [Gσ] represents a Gaussion smoothing operator at an outer scale [σ] and [Sx, Ss] denote the horizontal and vertical gradients calculated on an inner scale [ρ] in EPI. Optimal inner and outer parameters are found after testing a number of different parameter combinations, and the disparity can be computed using (3).

[J=Gσ*(SxSx)Gσ*(SxSs)Gσ*(SxSs)Gσ*(SsSs)=JxxJxsJxsJss ] (2)

[dx,s=?x?s=2JxsJss-Jxx ] (3)

To yield a higher quality depth map from the 4D light field, we need to add some optimization approaches. Since the local disparity estimation only takes into account the local structure in EPI, if there is no dominant orientation in the neighborhood, i.e. homogeneity or multiple orientations, the structure tensor cannot give a reliable estimation. Li [23] proposed a method to refine the certainty maps produced by the structure tensor. And the disparity map would be more convincing if applying consistent disparity labeling globally on the EPI [24]. However, this is a computationally very expensive procedure, and it is far too slow. In order to obtain quality disparity map, we employ denoising or globally optimization on individual views. Table 1 shows the speed and accuracy of depth estimation with optimization or not for this method compared to a multi?view stereo. It is clear that this method is much faster than the multi?view stereo because it does not need to match pixels at different locations. In addition, the accuracy of the depth map is a little better than the results using the multi?view stereo method since multi?view method is more likely to find the wrong matching patch, leading to lower accuracy.

3.2 View Synthesis Utilizing DIBR

After the step of disparity estimation, we synthesize arbitrary virtual views we need. View synthesis can be done by utilizing DIBR to generate new views with depth map available. Compared with the MPEG depth estimation which uses stereo matching based on 2 or 3 input views and view synthesis reference software, the results of the synthesized views show that light field method is better objectively for most of the scenes and always visually better [25]. The DIBR procedure is briefly introduced below.

For the camera is 1D parallel arrangement, once the depth value z is calculated, we can use DIBR easily by mapping every pixel from the reference view to the virtual view using (4) [26].

[xv=xr-f?tx,v-tx,rZ=xr-d ] (4)

where [xv] and [xr]are the positions of x in camera plane in virtual and real view respectively, and [d] is disparity which is related to the depth [z] and the baseline [(tx,v-tx,r)].

Applying the DIBR method, every pixel in reference view can be projected to the virtual view with disparity map according to (4). The geometry matching procedure is illustrated in Fig. 3. For a point C in virtual view, its unique corresponding point in reference view is A. If C is a point in integer pixel position, the intensity of C is equal to A. However, during the procedure of mapping, maybe more than one point in the reference view map to one point in the virtual view. To solve this problem, we can choose the corresponding pixel which has smallest depth in reference view. Another problem is that some positions in the virtual view cannot find their corresponding points in the reference views because of occlusion. Therefore, hole filling needs to be done, which has great impact on the quality of synthesized views. We can preprocess the depth data or inpaint the synthesizing view [27] to deal with this problem.

4 Conclusion

In this paper, we summarize several major methods of light field virtual view rendering based on EPI?representations in SMV application. Compared with stereo matching, the method of light field analysis can obtain better depth estimation. With a more accurate depth map, it is apparent that the virtual views quality is higher. In addition, rendering complexity is also included in this paper. Methods based on global line detection to estimate depth is more time?consuming than the methods based on local disparity estimation. Apparently, the local disparity estimation method saves more time at the cost of degrading the quality of virtual views. For current use cases, it is more suitable to use local disparity estimation to compute the depth map for SMV application because of the requirement of real?time.

References

[1] “Draft Call for Evidence on FTV,” ISO/IEC JTC1/SC29/WG11 MPEG2015/N15095, Feb. 2015.

[2] M. Tanimoto and T. Fujii, “FTV? Free Viewpoint Television,” ISO/IEC JTC1/SC29/WG11, M8595, Jul. 2002.

[3] “Use Cases and Requirements on Free?viewpoint Television (FTV),” ISO/IEC JTC1/SC29/WG11, Oct. 2015.

[4] M. Levoy and P. Hanrahan, “Light field rendering,” in Proc. 23rd ACM Annual Conference on Computer Graphics and Interactive Techniques, New Orleans, USA, 1996, pp. 31-42. doi: 10.1145/237170.237199.

[5] S. J. Gortler, R. Grzeszczuk, R. Szeliski, et al., “The lumigraph,” in Proc. 23rd ACM Annual Conference on Computer Graphics and Interactive Techniques, New Orleans, USA, 1996, pp. 43?54. doi: 10.1145/237170.237200.

[6] R. Szeliski, Computer Vision: Algorithms and Applications. Berlin Heidelberg, Germany: Springer Science & Business Media, 2010, pp. 628-628.

[7] S. Wanner and B. Goldluecke, “Variational light field analysis for disparity estimation and super?resolution,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 36, no. 3, pp. 606-619, 2014.

[8] S. Schwarz, R. Olsson, and M. Sjostrom, “Depth sensing for 3DTV: a survey,” MultiMedia, vol. 20, no. 4, pp. 10-17, 2013.

[9] G. Jiang, M. Yu, X. Ye, et al., “New method of ray?space interpolation for free viewpoint video,” in IEEE International Conference on Image Processing, Genova, Italy, 2005. doi: 10.1109/ICIP.2005.1530261.

[10] C. Kim, H. Zimmer, Y. Pritch Y, et al., “Scene reconstruction from high spatio?angular resolution light fields,” ACM Transactions on Graphics, vol. 32, no. 4, article no. 73. doi: 10.1145/2461912.2461926.

[11] T. Ishibashi, M. P. Tehrani, T. Fujii, et al., “FTV format using global view and depth map,” in IEEE Picture Coding Symposium (PCS), Krakow, Poland, 2012, pp. 29-32.

[12] L. Fan, X. Yu, Z. Shu, et al., “Multi?view object segmentation based on epipolar plane image analysis,” in IEEE International Symposium on Computer Science and Computational Technology, Shanghai, China, Dec. 2008, pp. 602-605.

[13] L. Jorissen, P. Goorts, S. Rogmans, et al., “Multi?camera epipolar plane image feature detection for robust view synthesis,” in IEEE 3DTV?Conference: The True Vision?Capture, Transmission and Display of 3D Video, Lisbon, Portugal, 2015, pp. 1-4.

[14] T. Ishibashi, M. P. Tehrani, T. Fujii, et al., “FTV format using global view and depth map,” in IEEE Picture Coding Symposium (PCS), Krakow, Poland, 2012, pp. 29-32.

[15] M. Matou?ek, T. Werner, and V. Hlavác, “Accurate correspondences from epipolar plane images,” in Proc. Computer Vision Winter Workshop, Bled, Slovenia, 2001, pp. 181-189.

[16] A. Criminisi, S. B. Kang, R. Swaminathan, et al., “Extracting layers and analyzing their specular properties using epipolar?plane?image analysis,” Computer Vision and Image Understanding, vol. 97, no. 1, pp. 51-85, 2005.

[17] R. C. Bolles, H. H. Baker, and D. H. Marimont, “Epipolar?plane image analysis: an approach to determining structure from motion,” International Journal of Computer Vision, vol. 1, no. 1, pp. 7-55, 1987.

[18] F. C. Calderon, C. Parra, and C. L. Ni?o, “Depth map estimation in light fields using an stereo?like taxonomy,” in IEEE XIX International Symposium on Image, Signal Processing and Artificial Vision, Armenia, Colombia, 2014, pp. 1-5.

[19] M. Uliyar, G. Putraya, S. V. Basavaraja, “Fast EPI based depth for plenoptic cameras,” in IEEE International Conference on Image Processing (ICIP), Melbourne, Australia, 2013, pp. 1-4.

[20] H. Lv, K. Gu, Y. Zhang, et al., “Light field depth estimation exploiting linear structure in EPI,” in IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Torino, Italy, 2015, 1-6.

[21] D. Dansereau and L. Bruton, “Gradient?based depth estimation from 4d light fields,” in Proc. IEEE International Symposium on Circuits and Systems, Vancouver, Canada, 2004, pp. 549-552. doi: 10.1109/ISCAS.2004.1328805.

[22] J. Luke, F. Rosa, J. Marichal, et al., “Depth from light fields analyzing 4D local structure,” Journal of Display Technology, vol. 11, no. 11, Nov. 2015. doi: 10.1109/JDT.2014.2360992.

[23] J. Li and Z. N. Li, “Continuous depth map reconstruction from light fields,” in IEEE International Conference on Multimedia and Expo, San Jose, USA, 2013, pp. 1-6.

[24] S. Wanner and B. Goldluecke, “Globally consistent depth labeling of 4D light fields,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, USA, 2012, pp. 41-48.

[25] L. Jorissen, P. Goorts, B. Bex, et al., “A qualitative comparison of MPEG view synthesis and light field rendering,” in IEEE 3DTV?Conference: The True Vision?Capture, Transmission and Display of 3D Video, Budapest, Hungary, 2014, pp. 1-4.

[26] D. Tian, P. L. Lai, P. Lopez, et al., “View synthesis techniques for 3D video,” in SPIE Optical Engineering Applications, International Society for Optics and Photonics, San Diego, USA, 2009, 74430T?74430T?11. doi:10.1117/12.829372.

[27] I. Daribo, H. Saito, R. Furukawa R, et al., “Hole filling for view synthesis,” in 3D?TV System with Depth?Image?Based Rendering. New York, USA: Springer, 2013, pp. 169-189. doi: 10.1007/978?1?4419?9964?1_6.

Manuscript received: 2015?11?15

主站蜘蛛池模板: 亚洲天堂久久久| 国产午夜精品一区二区三| 日本亚洲欧美在线| 国产色婷婷| 国产欧美日韩在线在线不卡视频| 成人毛片免费在线观看| 国产91视频免费观看| 欧美国产日韩另类| 久久一级电影| 亚洲日韩AV无码一区二区三区人| 欧美色99| 国产真实乱子伦视频播放| 欧洲欧美人成免费全部视频| a在线亚洲男人的天堂试看| 欧美全免费aaaaaa特黄在线| 亚洲国产欧美国产综合久久| 国产午夜福利在线小视频| 操美女免费网站| 亚洲免费人成影院| 露脸真实国语乱在线观看| 久久久受www免费人成| 国产精品理论片| 欧美日韩激情在线| 日韩无码精品人妻| 欧美激情第一欧美在线| 国产成人a在线观看视频| 性视频一区| 黄色国产在线| 日韩精品中文字幕一区三区| 免费在线色| 青青草久久伊人| 午夜天堂视频| 欧美激情福利| 久久免费视频6| 免费观看男人免费桶女人视频| 亚洲国产精品一区二区第一页免| 亚洲人视频在线观看| 自慰高潮喷白浆在线观看| 看av免费毛片手机播放| 高潮爽到爆的喷水女主播视频| 国产精品无码一二三视频| 欧美激情视频在线观看一区| 色偷偷男人的天堂亚洲av| 中文国产成人久久精品小说| 一区二区三区国产| 91欧美在线| 亚洲欧美综合在线观看| 欧美在线一二区| 久久精品91麻豆| 免费人成网站在线高清| 伊人激情综合网| 精品福利视频网| 3344在线观看无码| 暴力调教一区二区三区| 久久免费精品琪琪| 欧美成人精品高清在线下载| 最新亚洲av女人的天堂| 无码AV动漫| 欧美精品成人| 亚洲色图狠狠干| www中文字幕在线观看| 视频二区亚洲精品| 免费国产高清视频| 欧美福利在线观看| 最新亚洲人成无码网站欣赏网| 国产精品无码影视久久久久久久| 97一区二区在线播放| 国产aaaaa一级毛片| 国产H片无码不卡在线视频| 久久女人网| 国产一级毛片高清完整视频版| 不卡视频国产| 亚洲精品国偷自产在线91正片| 成人福利视频网| 国产一级片网址| 亚洲国产精品日韩专区AV| 成人a免费α片在线视频网站| 天天综合色天天综合网| 国产一区二区影院| 精品天海翼一区二区| 久久无码免费束人妻| 极品av一区二区|