Next Article in Journal
High-Precision Absolute Pose Sensing for Parallel Mechanisms
Next Article in Special Issue
Microsoft Azure Kinect Calibration for Three-Dimensional Dense Point Clouds and Reliable Skeletons
Previous Article in Journal
Missing and Corrupted Data Recovery in Wireless Sensor Networks Based on Weighted Robust Principal Component Analysis
Previous Article in Special Issue
Virtual Namesake Point Multi-Source Point Cloud Data Fusion Based on FPFH Feature Difference
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Noise-Resilient Depth Estimation for Light Field Images Using Focal Stack and FFT Analysis

School of Electrical and Data Engineering, University of Technology Sydney, Ultimo, NSW 2007, Australia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(5), 1993; https://doi.org/10.3390/s22051993
Submission received: 19 December 2021 / Revised: 23 February 2022 / Accepted: 1 March 2022 / Published: 3 March 2022
(This article belongs to the Special Issue Camera Calibration and 3D Reconstruction)

Abstract

:
Depth estimation for light field images is essential for applications such as light field image compression, reconstructing perspective views and 3D reconstruction. Previous depth map estimation approaches do not capture sharp transitions around object boundaries due to occlusions, making many of the current approaches unreliable at depth discontinuities. This is especially the case for light field images because the pixels do not exhibit photo-consistency in the presence of occlusions. In this paper, we propose an algorithm to estimate the depth map for light field images using depth from defocus. Our approach uses a small patch size of pixels in each focal stack image for comparing defocus cues, allowing the algorithm to generate sharper depth boundaries. Then, in contrast to existing approaches that use defocus cues for depth estimation, we use frequency domain analysis image similarity checking to generate the depth map. Processing in the frequency domain reduces the individual pixel errors that occur while directly comparing RGB images, making the algorithm more resilient to noise. The algorithm has been evaluated on both a synthetic image dataset and real-world images in the JPEG dataset. Experimental results demonstrate that our proposed algorithm outperforms state-of-the-art depth estimation techniques for light field images, particularly in case of noisy images.

1. Introduction

Depth estimation for planar images is essential for applications such as changing depth of focus, simulating subsurface scattering and shadow map**. Depth estimation is also crucial in computer vision applications such as robot vision, self-driving cars, surveillance and human–computer interactions. Depth estimation algorithms also play an important role in semantic segmentation algorithms [1,2]. To estimate depth maps, a wide range of stereo matching techniques have been proposed and implemented, and Scharstein and Szeliski present an in-depth analysis of these techniques [3]. In recent years, the introduction of light field images has made it possible to generate images at different focal lengths, extended depth of field and estimate the scene depth from a single image capture. Depth maps from light field images are essential for light field image compression techniques, reconstructing views from a sparse set of perspective views, increasing the number of perspective views and 3D reconstruction.
Compared to the techniques using a single image or stereo image pair for depth estimation, light field images enable researchers to explore techniques such as depth from correspondence and defocus from a single image. This is due to light field images capturing not only the pixel intensity but also the direction of the incident light with a single capture; however, the challenges in depth map estimation from planar images still apply, and are compounded by the challenges of estimating depth from light field images, including the presence of occlusion, textureless regions and overexposed regions in images.
Existing light field depth estimation techniques include stereo-based matching, using epipolar images, depth from defocus and more recent techniques that use neural networks. Stereo matching for depth estimation suffers from ambiguity due to partial occlusions [4]. Since the stereo pair are images, the information lost by the occlusions cannot be recovered but can only be approximated using a smoothness term to fill the gaps. In contrast, epipolar images are formed by stacking together the sub-aperture images in both the horizontal and vertical direction as shown in Figure 1b,c, respectively. A slice through this 4D block reveals the depth of the pixels, as the slope of the line reflects the depth information as shown by the red, green and blue parallelograms in Figure 1. The pixels in the central view that do not move with changing sub-aperture views have a zero slope and are seen as a straight line as shown by the blue parallelograms in Figure 1b. The pixels that are closer to the camera incline to the right as shown by the green parallelograms in Figure 1c, and the pixels that are further away from the camera incline to the left as shown by the red parallelograms in Figure 1. However, basic line fitting techniques to estimate depth would not give robust results and the reconstructions are generally noisy [5]. Schechner and Kiryati [6] have extensively studied depth from defocus and correspondence techniques and have compared the advantages and disadvantages of both cues. Finally, Convolutional Neural Networks (CNN) are a known approach in imaging applications such as object recognition, human–computer interaction, image segmentation and stereo depth estimation [7]. However, the lack of training data in existing light field images datasets makes it hard to train and test the network for light field images [8].
In this paper, we exploit the characteristic of light field images having multiple focal planes that are captured in a single image to generate the disparity or depth map. Our algorithm is designed to use the concept of depth from defocus by a one-to-one comparison between the focal stack image and the central all-in-focus image. We show that our approach is noise resilient in depth estimation, and out-performs the current state-of-the-art benchmark algorithms in the presence of noise. The proposed algorithm is compared to four benchmark techniques from Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for synthetic images with and without noise. All the four techniques discuss their algorithms’ noise resilience in their work. Strecke et al. [9] suggest mixing the focal stack cost volume with a stereo correspondence cost volume to increase resilience against noise. Wang et al. [10] validate the robustness of their system on noisy image, and report the F-measure values of each algorithm. Zhang et al. [11] suggest reducing the number of bins for depth estimation in the presence of noise, but do not compare their algorithm’s noise performance result to other algorithms. The multi-stream CNN proposed in Shin et al. [12] uses 2 × 2 kernel that reduces the effect of noise.

2. Our Contribution

The main contributions of our work are:
  • To reduce the dependence of depth accuracy on RGB values of individual pixels compared in the image patches, we propose a method that uses frequency domain analysis to estimate the depth map for light field images.
  • The key contribution of our approach is noise resilience in depth estimation. Our analysis confirms the hypothesis that comparing focal stack image patches in the frequency domain improves depth map accuracy, especially in the presence of noise. We shown that our algorithm out-performs the current state-of-the-art benchmark algorithms in the presence of noise.

3. Background

Conventional photography is only able to capture limited information from the light passing through the camera lens. In general, cameras record the sum of the intensities of light rays striking each point in the image and not the total amount of incident light traveling along different rays that contribute to the image [13]. In contrast, light field imaging technology captures rich visual information by representing the distribution of light in free space [14], which means that a light field image captures the pixel intensity and the direction of the incident light. The additional dimensions of data captured enables the generation of images at different focal lengths and extended depth of field using ray-tracing techniques. This allows for image manipulation in a more flexible way [15]. Capturing a light field can be a challenging task as a light field image captures the pixel intensity and the direction of the incident light. The two techniques that can be used to capture a light field image are: first, using an array of video cameras as described in [16]; then second, using a plenoptic camera as shown in Figure 2. Once the light field image is captured, raytracing can be applied to filter out the light rays that do not contribute to a particular depth. Rearranging the rays then estimates where the light rays would terminate if the camera was focused on the desired depth [17].
The model that describes the distribution of light is known as a plenoptic function. The plenoptic function describes light as a function of position, angle of incidence, wavelength and time [14]. As shown in Figure 3, the most common way to represent the 4D plenoptic function is to parameterize the light rays as an intersection of the light ray on two planes placed at arbitrary positions [14]. The plenoptic function can thus be defined as L (u, v, s, t); where (u, v) and (s, t) denote the position of intersection of the light ray on the two planes, respectively. The function L (u, v, s, t) is a two-plane model of light field in which the st plane can be considered as a set of cameras that have the focal plane on the uv plane [14].
Finally, a light field image can be visualized by representing the image as an Epipolar Plane Image (EPI). The EPI can be obtained by fixing one coordinate in both spatial and angular domains, while plotting pixel intensities as the other coordinates of the spatial and angular domain are varied [11], as shown in Figure 4c.

4. Related Work

Current key approaches to light field depth estimation techniques include stereo based matching, using epipolar images, depth from defocus and, more recently, the use of Convolutional Neural Networks (CNNs).

4.1. Depth Estimation Using Stereo Matching

Many of the proposed stereo matching algorithms are based on graph cuts and energy minimization techniques by using different constraints. Kolmogorov and Zabih [19] and Woodford et al. [20] combine visibility and smoothness terms for effective optimization Bleyer et al. [21], on the other hand, consider the pixel appearance, global MDL constraint, smoothing, soft segmentation, surface curvature and occlusion. However, the stereo correspondence methods suffer from ambiguities while dealing with noisy and aliased regions, and the narrow baseline for real camera LF images makes it difficult for these algorithms to solve the occlusion problem [11].

4.2. Depth Estimation Using Epipolar Plane Images

Depth estimation using Epipolar Plane Images (EPI) is another popular approach. Each slice through the 4D light field representation can be used to analyze the line structure to estimate the depth of the pixel under inspection as shown in Figure 1a,b. Johannsen et al. [22] propose a method that learns the structural information from the central view, and compare them to the 2D 5 × 5 epipolar image patch to estimate the depth map. Wanner and Goldluecke [23] address the problem of reflective surfaces in the image that cause irregularities in the epipolar plane. They propose the use of higher order structure tensors to overcome this problem. Criminisi et al. [24] use the epipolar images to build and minimize the matching cost for each pixel that accounts for the pixel intensity value, gradient pixel value and spatial consistency. Criminisi et al. [24] use canny edge operator, while Wanner and Goldluecke [25] use structure tensor to obtain local disparity estimates by computing the weighted mean of each side of the line in the EPIs and then finding the distance between the two means [11]; however, occlusions and noise could cause the pixels on both sides of the lines to be different. Zhang et al. [11] propose a spinning parallelogram operator that tries to estimate the local direction of a line at a specific point in an EPI, and to avoid problem of the inconsistency of pixel distribution on either sides of the line, they consider each side separately. This assumption makes their depth estimation algorithm more robust to occlusions and noise, then Criminisi et al. [24] and Wanner and Goldluecke [25]’s algorithm. We compare our results with Zhang et al. [11] for noisy images and their EPI-based technique generates outliers that severely effects the accuracy of their depth maps.

4.3. Depth Estimation Using Defocus

Depth estimation from defocus has been studied extensively e.g., Schechner and Kiryati [6], who compare the advantages and disadvantages of using defocus or correspondence cues. Research on depth from defocus also extends beyond using LF images, estimating the depth map using a single image [26,27]. Nayar and Nakagawa [28] in their work demonstrate that the defocused imaging system plays the role of a low-pass filter to estimate depth using the measure of focus between image points. Strecke et al. [9] create four separate stacks using only the views right of, left of, above and below the reference view. The assumption is that the baseline is small enough so that if occlusion is present it occurs only in one direction of view point shift. Tao et al. [29] proposed to combine the defocus and correspondence cues using a MRF as the global optimization process. Building on the concept of combining defocus and correspondence cues for depth estimation [29], Wang et al. [10] take into account occlusion to estimate a more accurate depth map. To account for occlusion, an angular patch is generated for the reference pixel to then check for the photo-consistency of that patch. However, similar to other proposed techniques, the algorithm does not perform well in regions of low texture [10].

4.4. Depth Estimation Using Convolutional Neural Networks

Convolutional Neural Networks (CNN) have been used extensively in image processing applications such as object recognition, human–computer interaction, image segmentation and stereo depth estimation [7]. Over the past few years, research has been conducted to use CNN for depth estimation in light field images [7,8,30]. The main concern in using CNN for light field images is that the existing light field datasets are insufficient in size to train the network, and datasets do not have accurate ground truth depth included [30]. To address this problem, Heber and Pock [30] generated a synthetic light field dataset using the raytracing software POV-Ray, Feng et al. [7] use the Lytro camera to capture the light field images and then use a 3dMD scanner [31] to capture the ground truth, Shin et al. [12] on the other hand augment the data through scaling, center view change, rotation, transpose and color that are suitable for light field images. In their approach, Heber and Pock [30] extract the information from the vertical and horizontal EPIs to input into the CNN. The resultant depth map is optimized by solving a convex energy minimization problem. Unlike Heber and Pock [30] that use only one directional EPI, Shin et al. [12] construct a multi-stream networks for four viewpoints, horizontal, vertical, left and right diagonal directions. They show that their multi-stream network reduces the reconstruction error over single-stream network by 10%. They claim that the use 2 × 2 kernel in the algorithm reduces the effect of noise. In this paper, we compare our results with their multi-stream network on both synthetic images with and without noise, and even though their algorithm out-performs our algorithm for noiseless images, our algorithm out-performs their results for all images with added noise.

5. Methodology

The methodology presented in this paper exploits the characteristic of light field images in having multiple focal planes that are captured in a single image. The flow of the algorithm is represented in Figure 5. As shown in Figure 5, the methodology can be divided into four main sections: initial depth estimation, focal stack generation, patch generation and comparison and depth refinement. Unlike the depth from defocus algorithms described in the literature review, our approach uses frequency patch analysis, which makes the algorithm more resilient to noise.

5.1. Initial Depth Estimation

The initial depth estimation algorithm only determines the maximum and minimum depth values, rather than the intermediate depth values. The depth or slope difference between two consecutive focal stack images for the initial depth estimation algorithm is taken to be 0.2; experimental evaluation proved that this value sufficiently covers the depth range to permit accurate estimation of periphery depth values. This initial depth estimation stage improves the depth estimation in two ways: firstly, it reduces the computational time as only the relevant focal stack images are generated; secondly, reduces the number of redundant images to then also reduce the possibility of misdetections.

5.2. Focal Stack Generation and Image Pre-Processing

A single LF image can be used to generate an all-in-focus image and also to generate the same scene at different focal lengths and narrow depth of field. The sub-aperture images can be obtained by holding (u,v) fixed and considering all (s,t) [17]. The central sub-aperture image is the all-in-focus image. Figure 6 shows the conceptual model of a light field image when refocusing at a virtual plane (s′,t′). Thus, to refocus the image to a different focal plane (s′,t′), the shifted versions of the sub-aperture images are summed together.
Following the same concept, the focal stack is generated using the shift-sum filter, which is a depth-selective filter that functions like a planar focus. The filter works by shifting (u,v) slices of the LF image to a common depth and then adding the slices together to obtain the 2D image. The value of the slope controls the amount of shift that corresponds to the image being focused on a different focal plane. The difference in slope values between two consecutive images in the focal stack for our work is 0.01, which implies that to obtain an accurately refocused image at that small depth difference the sub-aperture images have to be shifted at multiples of 0.01 pixels. To shift the sub-aperture images with subpixel accuracy we use a frequency domain approach. In this approach, the relationship between the image shift in the spatial and frequency domains is shown in Equation (1), where s0 and t0 is the slope value and u, v is the sub-aperture location. The amount by which the sup-aperture image has to be shifted to refocus at a particular depth is the product of s0, u and t0, v.
f ( s + s 0 , t + t 0 ) = F ( u , v ) e j 2 π ( u s 0 + v t 0 N )
The refocused images can then be generated by either averaging the shifted sub-aperture pixels or by using the median value of the pixels. Figure 7 shows a comparison of the blur around the depth discontinuities when focusing on the background for both the summing techniques. It is clear from the magnified parts shown in Figure 7a–c that the amount of defocus blur around the depth discontinuities using the averaged pixel value is very large compared to the amount of defocus blur using the median value.
The technique used in this paper estimates the depth map by matching image patches from the focal stack to the central sub-aperture image, thus it must be ensured that all the edges and textured regions in the image are well defined in both the central all-in-focus image as well as the focal stack images to minimize the number of misdetections. This problem is addressed by adding the gradient of the image to itself. The main advantage of adding the gradient relies on the fact that in a refocused focal stack images, the textured regions in the image that are in focus maximally contribute to the gradient image, while the out-of-focus objects will contribute the least. This pre-processing step ensures that the object boundaries and textured regions are exaggerated in the focal stack images drastically reducing the number of misdetected patches, in turn reducing the dependence on the post-processing steps. As the purpose of the gradient addition step is only to enhance the textured regions and boundaries on the focal stack images that are in-focus, unless the shadows cause the region to become textureless in the image, our algorithm is not affected by this step. The comparison between the accuracy of the estimated depth map with and without the gradient addition is shown in Figure 8 and it can be seen in Figure 8c,f that the part of the image with shadows are also estimated accurately.

5.3. Patch Generation and Comparison

The focal stack images that are acquired in the previous stage are divided into smaller image patches and then those individual patches are compared with the corresponding block in the all-in-focus image. Since the accuracy of the approach is dependent on the similarity check of the individual image patches, it is crucial that a block of appropriate shape and size is selected. Initial tests for patch selection were performed with a square patch of size 10 × 10 pixels. This was the preliminary test that was performed to validate the approach. The test was then repeated with patches of different sizes. The results showed that smaller window sizes covered the image regions and boundaries more accurately. As the window size decreases, the processing time increases as the number of patches that are being compared increase; however, the nature of square-shaped patches over or under-compensated object boundaries that were slanted or curved in the image.
In testing cross shape patches, the area that is uncovered in the gaps between four consecutive cross is less than that of the primary cross window size. This meant that to cover the entire image without any gaps, cross patches of two different sizes were used. The shape and size of the primary cross governs the shape and size of the secondary cross as shown in Figure 9. To address misdetections due to small window size using cross patches, an overlap** window is used. For the proposed algorithm, we use the two cross patches of size 4 × 4 pixels as shown in Figure 9. In Figure 9 the red and green squares are the pixels that are considered for matching with the all-in-focus image patch, but only the pixels highlighted in red in the red square and the pixels highlighted in green in the green square are used to generate the depth map. The patch size can be reduced lower than 4 × 4 pixels, although experimental tests revealed that using a patch of size smaller than 4 × 4 pixels does not improve the depth map accuracy and increases the computational time.
By comparing the FFT of the image patches, we are no longer looking at individual pixel values when comparing the image patches, but a frequency domain representation of those patches, which makes the comparison more robust to noise. To illustrate the proposed approach Figure 10 shows the central sub-aperture image of the LF image, a 4 × 4 pixel patch taken from the image and the FFT of the patch generated. The focal length for synthetic images lies between a slope of −4 to +4 and for a Lytro camera, the image lies between slopes of −2 to +2. We therefore correspondingly generated the focal stack from a slope of −4 to +4 for synthetic LF images, and from the slope of −2 to +2 for real LF images. The slope interval between two consecutive focal stack images defines the number of depth levels in the depth map. We found through experimentation that for the purposes of our work, the slope can be varied at an interval of 0.01, as using an interval below 0.01 does not show any significant change in the focus for consecutive focal stack images for both the synthetic and real LF images. Thus, for a synthetic LF image the depth map can have up to 801 depth levels as the focal stack is generated from a slope of −4 to +4 at a slope interval of 0.01, while for a real LF image the depth map can have up to 401 depth levels as the focal stack is generated from a slope of −2 to +2 at a slope interval of 0.01. The depth levels for the depth map can be increased by reducing the slope interval between the focal stack images, although as the depth levels increase the computation time also increases. For visual brevity in Figure 11, only 8 refocused images are considered from −4 to +4 slope at an interval of 1. It is clearly seen that the fifth patch in Figure 11 is the most similar to the reference patch.

5.4. Depth Map Refinement

The depth map evaluated up to this stage has a few patches that are not detected correctly, and since the patches are shaped as a cross, it creates a depth map with jagged edges, thus the object boundaries need to be refined in order to restore the shape of the object. Figure 12 and Figure 13 show the comparison between the ground truth and the estimated depth map before and after this refinement step for the synthetic and real images. The disparity map is refined in two steps using an iterative approach. Firstly, the histogram of the disparity map is checked for the number of pixels that are present at each depth. If the number of pixels at a particular depth falls below the threshold value, those pixels are filled with the maximum likelihood value of the pixels in the depth map at that position using the pixel value that occurs the most times in the neighboring pixels. The second step is similar to the first step, but instead of looking at individual pixels, the cross patches are considered. This step checks for any isolated patches in the image that have different surrounding depth patches. Once these patches are isolated, the patch is filled with the value of pixels with maximum likelihood in the depth map at that patch position using the pixel value that occurs the most times in the neighboring pixels.

6. Misdetection Analysis

The depth estimation in the proposed algorithm compares the FFT of the all-in-focus patch and the patches at different focal positions in the focal stack to check for the least error to then accordingly estimate the depth. The quantitative results for ‘dot’ image presented in Table 1 confirm that both techniques give comparable results. The advantage to using the FFT domain over the spatial RGB patch is that the number of misdetections is drastically reduced with the FFT, which reduces the algorithm dependence on the depth map refinement stage. A closer visual comparison in Figure 14 also shows that the depth boundaries are sharper and the depths are more accurately represented for the results using FFT. The ‘dot’ image is one of the more challenging images from the dataset due to the added Gaussian noise to approximate thermal and shot noise [32], and the shape and size of the objects in the image. In further evaluation, image-level comparisons for the proposed algorithm are shown in Figure 14. It can be seen that the noise in the image considerably reduces the depth map accuracy when using RGB patches, where parts of objects in the image are completely misdetected; whereas, for the depth map generated by using the proposed FFT comparison, the results are more noise-resilient. The proposed algorithm also outperforms the state-of-the-art for the ‘dot’ image, as can been seen in Figure 15.
Figure 16 shows central view, ground-truth depth map and the estimated depth map for the Rosemary image in the synthetic image dataset. Our algorithm produces an inaccurate depth map for the Rosemary image with a Badpix 0.07 value of 0.34. The error is caused because the wall in the background and the vase in the foreground have a smooth and textureless surface, which make the two indistinguishable by our algorithm. It is important to note that the misdetection is not caused by the shadows in the image as the carpet at the bottom of the image is not misdetected even though shadows falls on the carpet as well, and the cotton image in Figure 8 also show that shadows do not effect the depth map accuracy for our algorithm.
Figure 17 shows the mean squared error (MSE) for the patch that is estimated to the correct depth at different focal lengths with the patch of the all-in-focus image. This trend in the graph indicates that the image patch is depicting the correct depth: in the example shown in Figure 17, the plot takes a significant dip as it reaches the least MSE, which is the true depth for that particular patch. As the patch goes further away from the correct depth value, the graph makes a similar curve on both the sides of the focal stack. Even though in our work we are only using the patches with the least MSE to estimate the depth map, it is an important observation to see that the graph almost traces a bell curve. This shows that the MSE value is similar when defocusing toward or away from the patch in focus. In contrast, Figure 18 shows an example of a patch being misdetected in the ‘cotton’ image. Although the graph follows a similar trend to Figure 17, the graph has two considerable dips, one at the correct depth of slope 1.2, and the other at the incorrect depth of slope −1.6. It is also important to note that the number of misdetections are less even though the depth map has not been refined at this stage.

7. Experimental Results

The results of the proposed algorithm were evaluated on both real and synthetic light field image datasets. For the synthetic data, the 4D light field dataset [32] was used. The dataset is widely used for validations of depth estimation algorithms for light field images as it contains ground-truth disparity and depth maps. The dataset contains 28 image with a 9 × 9 sub-aperture images with a resolution of 512 × 512. For evaluation of our algorithms with the benchmark algorithms we have selected 10 images as each image contains different materials, lighting conditions and complex structures. Figure 19a–c and Figure 20d have finer detail and complex occlusions. Figure 19c,d have transparent and reflective surfaces. Figure 19d and Figure 20a,e have shadows. Figure 19b and Figure 20b,c are abstract scenes with complex textures. For the real data the EPFL light field dataset [33] was used. The real image dataset contains 138 LF images in LFR (light field raw) file format captured by a Lytro Illum camera with a 15 × 15 sub-aperture images with a resolution of 434 × 625. The Lytro Illum camera used for LF image acquisition have different calibration data and the LFR files were processed by their corresponding calibration data. Figure 21a,c and Figure 22a–c contain finer objects and complex surfaces such as perforated metal and fences. Figure 21b and Figure 22c,d contain textureless or overexposed regions such as the sky. Figure 21a,b and Figure 22d,e show a gradual change in depth and Figure 21e contains complex structures such as the branches and trees.
The depth maps generated by our proposed approach initially calculates the depth range, and the depth levels vary according to the maximum and minimum depth values present in each LF image. The slope for real data is within the range of +2 to −1, whereas for the synthetic data lies within the range of −1 to +4 and the slope interval used for both type of images is 0.01, as explained in Section 5.2 and Section 5.3. The number of depth levels can be increased or decreased by reducing or increasing the slope interval between focal stack images. The proposed algorithm is compared to four benchmark techniques from Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] using the BadPix metric that specifies the percentage of pixels where disparity deviates by less than 0.07, 0.03 and 0.01 pixels from the ground truth. We have chosen these four techniques are they are the state-of-the-art for the different depth estimation techniques. Strecke et al. [9] and Wang et al. [10] use depth from defocus, Zhang et al. [11] use EPIs, whereas Shin et al. [12] use CNN for depth estimation. In order to compare the depth maps using different algorithms to the ground truth, all output disparity maps are normalized to the ground truth depth map range. For Strecke et al. [9] and Shin et al. [12], normalized results are directly compared to the ground-truth disparity map. For Wang et al. [10] and Zhang et al. [11] the disparity map is normalized before comparing it to the ground truth.

7.1. Synthetic LF Images

The LF images in the 4D Light Field Dataset [32] comprise 9 × 9 sub-aperture images. For the synthetic images, the images in Figure 19 and Figure 20 show the error pixels in red where depth deviates by more than 0.07 from the ground truth and the pixels in green where depth deviates by less than 0.07. Table 2 and Table 3 compares the ground truth images to the disparity maps generated by the algorithms that are being tested using the Badpix metric. The LF images generated synthetically have little to no noise compared to the real LF images, and thus the estimated depth maps have less misdetections and the depth boundaries are well defined on the synthetic data compared to the real data. Table 2 and Table 3, which are a comparison of the depth maps with the ground-truth depth map, show that the proposed algorithm out-performs the state-of-the-art algorithms in the two criteria of Badpix 0.07 and 0.03 for the ‘dots’ images from the synthetic images in the dataset. On visual inspection of Figure 19 and Figure 20, it shows that even though the noise level is increased in the bottom part of the image, the background is region is still detected accurately. For ‘medieval 2’ image, the region in the image near the window on the top left and near the door on the bottom right has a dark shadow, which is the common area misdetected for all algorithms. Shin et al. [12]’s produce the least errors around object boundary for the synthetic light field images. The ‘kitchen’ and ‘museum’ image in Figure 19c,e shows how the error pixels are in the same regions for all the estimated depth maps. The reason for the similarity is those regions in the image are either transparent or reflective surfaces. Shin et al. [12] show fewer errors around these regions as they explicitly use a mask for these type of surfaces while training their network. The depth map for our algorithm, Shin et al. [12] and Strecke et al. [9] give similar results for the background and foreground region in the ‘kitchen’ and ‘museum’ image, whereas the depth maps from Wang et al. [10] and Zhang et al. [11] produce errors in the background and foreground regions. For the ‘museum’ image in Figure 19e, and the ‘pillow’, ‘platonic’ images in Figure 20a,b, our proposed algorithm out-performs the non-CNN based algorithms at Badpix 0.03, with comparable results for the other two criteria. Out of the three images mentioned above, for the ‘museum’ image the main reason for the errors is the reflective display case and the bright display case lighting. For the ‘platonic’, ‘pyramids’ and ‘tomb’ images in Figure 20, our depth map generates error only at depth boundaries and all other regions are estimated accurately and is comparable to Shin et al. [12]’s CNN approach. Shin et al. [12] produce high accuracy depth maps and are also able to distinguish accurate depths for occlusions and smaller objects in the image, but the accuracy reduces for transparent or reflective surfaces and for noisy images.

7.2. Real LF Images

The proposed algorithm is not able to distinguish objects in the image that are less than 4 × 4 pixels in width due to the patch size used, but using patches of size less than 4 × 4 pixels drastically increases the number of misdetected depth patches and also increases the computational time. The image result displayed in Figure 21 and Figure 22 on visual inspection shows similar outcomes as for the synthetic images. The central view for four images in the Figure 21a,b and Figure 22d,e show a gradual change in depth and the depth maps correspondingly show the gradient change. For the proposed algorithm, the images in Figure 21a,c and Figure 22c with the chain fences, the regions where the chain has a shadow cast over it is misdetected. The chain fences in all three images for all the algorithms have been under or over-compensated. The lorikeet image in Figure 22e is a complex image with the leaves and branches, but the proposed algorithm performs comparatively with Strecke et al. [9] and Zhang et al. [11]; on the other hand, with Wang et al. [10] major parts of the image are misdetected. For the ‘perforated metal 1’ image Figure 22a, parts of the image in the far background and foreground are represented with less error as compared to all the other depth maps, whereas in the ‘perforated metal 3’ image Figure 22b, Wang et al. [10] better estimates the depth around the holes in the metal frame. In Figure 21 and Figure 22, misdetections can been seen for textureless regions for the ‘backlight 1’ image Figure 21b, the ‘perforated metal’ image Figure 22a, and the ‘university’ image Figure 22d, i.e., the regions of the image with the sky.

7.3. Noisy Image Analysis

The results in Figure 15 and Figure 23 and Table 4 demonstrate that the proposed algorithm is more noise-resilient than existing approaches. To further explore our algorithm’s noise resilience, Gaussian noise approximates thermal and shot noise in images, and Gaussian noise with zero mean and variance of 0.01 was added to the 4D Light Field Dataset [32]. Gaussian noise was chosen as it approximates thermal and shot noise in images, as at large light levels, the Poisson distribution that describes shot noise approaches a normal distribution and can be approximated using Gaussian noise.
The proposed algorithm out-performs the benchmark algorithms for all images in all three criteria as shown in Table 5 and Table 6, Figure 24 and Figure 25. The algorithm from Zhang et al. [11] and Wang et al. [10] generates a disparity map that uses the maximum and minimum depth values from the ground-truth depth map to scale the disparity map accordingly. The problem with noisy images is that misdetections and outliers make the re-scaling unreliable and nontrivial. It is clear from the results shown in Table 5 and Table 6 that the accuracy of the proposed algorithm is also significantly affected for most of the images, but the algorithm is still able to estimate a depth map with comparatively high accuracy compared to the state-of-the-art algorithms. The average accuracy for our algorithm for Badpix 0.07 across the ten images used for testing shown in Figure 24 and Figure 25 is 0.75, whereas for Shin et al. [12], Strecke et al. [9], Wang et al. [10] and Zhang et al. [11] it is 0.6, 0.3, 0.26 and 0.11, respectively. For the ‘dots’, and ‘pyramids’ images the accuracy is over 95% out-performing the state-of-the-art algorithms. The images with finer details such as the ‘backgammon’ image in Figure 24a has a considerable amount of misdetections, but algorithm is still able to obtain a Badpix 0.07 value of 0.74. The ‘kitchen’ and ‘museum’ image in Figure 24c,e shows errors for transparent or reflective surfaces, but the depth map for our algorithm estimates the background and foreground region in the image accurately with Badpix 0.07 value of 0.63 and 0.69, whereas the depth map from Wang et al. [10] and Zhang et al. [11] produces errors in the background and foreground regions and Badpix 0.07 value below 0.25. Shin et al. [12], on the other hand, shows an Badpix 0.07 value 0.5 and 0.52, but large parts of the background and foreground region are misdetected. An important observation that can be drawn from Table 2, Table 3, Table 5 and Table 6 is that without the added noise the average accuracy for the proposed algorithm, Shin et al. [12], Strecke et al. [9], Wang et al. [10] and Zhang et al. [11] is 0.96, 0.9, 0.89, 0.79 and 0.86, respectively. After the noise is added to the images, our accuracy reduces to 0.75, whereas for Shin et al. [12] it reduces to 0.6, Strecke et al. [9] it reduces to 0.3, Wang et al. [10] it reduces to 0.26 and Zhang et al. [11] it reduces to 0.11.

7.4. Runtime Complexity Analysis

The code for the proposed algorithm was implemented in MATLABTM on an Intel i7 machine at 1.9 GHz and 16 GB of RAM. The results for Strecke et al. [9], Wang et al. [10] and Zhang et al. [11] were generated using the same machine. Table 7 and Table 8 show the average runtime for synthetic images and real images, respectively, over all the images in the dataset for a single run for the proposed algorithm compared to the state-of-the-art algorithms. The runtime for Strecke et al. [9] for real images is lesser than for synthetic images, as running their complete code requires the parameter files that are not available for the real images used. As Shin et al. [12] propose a CNN approach, their network has to be trained and their network takes 5 days to train.
The runtime for the proposed algorithm varies for each of the images as the algorithm first calculates the maximum and minimum depths for the image and then generates the focal stack. The runtime for the algorithm can be divided into four stages: the first stage calculates the initial depth, which takes an average of 52.18 s; all of the pre-processing steps in the second stage for the images takes an average of 0.21 s per image; the third stage generates the focal stack where majority of the processing time is spent, an average duration of 2.17 s per image; and the last stage is where the depth map is estimated and refined, which on average is 0.8 s per image. For real images the initial depth estimation stage takes an average of 150 s; all the pre-processing steps for the images takes an average of 0.21 s per image; the third stage generates the focal stack where majority of the processing time is spent, an average duration of 9.68 s image, and the last stage estimating and refining the depth map on average takes 1.02 s per image. The run time can be approximated by adding the per image times for each stage and multiplying it by the number of focal stack images. The runtime for the real images is more than the synthetic images as the number of sub-aperture images is more in the case of real images.

8. Conclusions

Depth maps from light field images are essential for light field image compression techniques, reconstructing views from a sparse set of perspective views, increasing the number of perspective views and 3D reconstruction. In this paper, we proposed a depth estimation algorithm that works on the concept of depth from defocus. Our experimental evaluation shows that the proposed algorithm outperforms state-of-the-art algorithms for both synthetic and the real light field image datasets. We also show that comparing the FFT of the images patches instead of the RGB patch directly increases our algorithm’s resilience to noise.
One key advantage of our algorithm is that it can be used to estimate the depth map using focal stack images captured by a 2D camera. The limitation of the proposed methodology is that the size of the patches makes it difficult to distinguish objects smaller than 4 × 4 pixels. To address this limitation we are currently investigating how the algorithm would perform if we would increase the resolution of the LF image or the focal stack images.

Author Contributions

Conceptualization, R.S., S.P. and E.C.; methodology, R.S.; software, R.S.; validation, R.S., S.P. and E.C.; formal analysis, R.S.; investigation, R.S.; resources, R.S.; data curation, R.S.; writing—original draft preparation, R.S.; writing—review and editing, R.S., S.P. and E.C.; visualization, R.S.; supervision, S.P. and E.C.; and R.S. drafted the manuscript, which was revised by all authors. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Australian Government Research Training Program & University of Technology Sydney.

Data Availability Statement

The results presented in the paper are available at https://github.com/rishabhsharma27/Depth_estimation_results, accessed on 23 December 2021.

Acknowledgments

The author would like to thank his supervisory panel for their encouragement and support. The authors would also like to thank the anonymous reviewers for their very knowledgeable comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhou, J.; Yang, D.; Cui, Z.; Wang, S.; Sheng, H. LRFNet: An Occlusion Robust Fusion Network for Semantic Segmentation with Light Field. In Proceedings of the 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), Virtual, 1–3 November 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1168–1178. [Google Scholar]
  2. Hu, X.; Yang, K.; Fei, L.; Wang, K. Acnet: Attention based network to exploit complementary features for rgbd semantic segmentation. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1440–1444. [Google Scholar]
  3. Scharstein, D.; Szeliski, R. A taxonomy and evaluation of dense two-frame stereo correspondence algorithms. Int. J. Comput. Vis. 2002, 47, 7–42. [Google Scholar] [CrossRef]
  4. Chen, C.; Lin, H.; Yu, Z.; Bing Kang, S.; Yu, J. Light field stereo matching using bilateral statistics of surface cameras. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1518–1525. [Google Scholar]
  5. Zhang, Y.; Lv, H.; Liu, Y.; Wang, H.; Wang, X.; Huang, Q.; **ang, X.; Dai, Q. Light-Field Depth Estimation via Epipolar Plane Image Analysis and Locally Linear Embedding. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 739–747. [Google Scholar] [CrossRef]
  6. Schechner, Y.Y.; Kiryati, N. Depth from Defocus vs. Stereo: How Different Really Are They? Int. J. Comput. Vis. 2000, 39, 141–162. [Google Scholar] [CrossRef]
  7. Feng, M.; Wang, Y.; Liu, J.; Zhang, L.; Zaki, H.F.M.; Mian, A. Benchmark Data Set and Method for Depth Estimation from Light Field Images. IEEE Trans. Image Process. 2018, 27, 3586–3598. [Google Scholar] [CrossRef] [PubMed]
  8. Heber, S.; Yu, W.; Pock, T. Neural EPI-Volume Networks for Shape from Light Field. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2271–2279. [Google Scholar] [CrossRef]
  9. Strecke, M.; Alperovich, A.; Goldluecke, B. Accurate depth and normal maps from occlusion-aware focal stack symmetry. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 22–25 July 2017; pp. 2814–2822. [Google Scholar]
  10. Wang, T.; Efros, A.A.; Ramamoorthi, R. Occlusion-Aware Depth Estimation Using Light-Field Cameras. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 3487–3495. [Google Scholar] [CrossRef]
  11. Zhang, S.; Sheng, H.; Li, C.; Zhang, J.; **ong, Z. Robust depth estimation for light field via spinning parallelogram operator. Comput. Vis. Image Underst. 2016, 145, 148–159. [Google Scholar] [CrossRef]
  12. Shin, C.; Jeon, H.G.; Yoon, Y.; Kweon, I.S.; Kim, S.J. Epinet: A fully-convolutional neural network using epipolar geometry for depth from light field images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4748–4757. [Google Scholar]
  13. Ng, R. Digital Light Field Photography; Stanford University: Stanford, CA, USA, 2006. [Google Scholar]
  14. Wu, G.; Masia, B.; Jarabo, A.; Zhang, Y.; Wang, L.; Dai, Q.; Chai, T.; Liu, Y. Light field image processing: An overview. IEEE J. Sel. Top. Signal Process. 2017, 11, 926–954. [Google Scholar] [CrossRef] [Green Version]
  15. Mousnier, A.; Vural, E.; Guillemot, C. Partial light field tomographic reconstruction from a fixed-camera focal stack. ar**v 2015, ar**v:1503.01903. [Google Scholar]
  16. Wilburn, B.; Joshi, N.; Vaish, V.; Talvala, E.V.; Antunez, E.; Barth, A.; Adams, A.; Horowitz, M.; Levoy, M. High performance imaging using large camera arrays. ACM Trans. Graph. TOG 2005, 24, 765–776. [Google Scholar] [CrossRef] [Green Version]
  17. Ng, R.; Levoy, M.; Brédif, M.; Duval, G.; Horowitz, M.; Hanrahan, P. Light field photography with a hand-held plenoptic camera. Comput. Sci. Tech. Rep. CSTR 2005, 2, 1–11. [Google Scholar]
  18. Levoy, M.; Hanrahan, P. Light field rendering. In Proceedings of the 23rd annual Conference on Computer Graphics and Interactive Techniques, ACM, Virtual Event, 9–13 August 2021; pp. 31–42. [Google Scholar]
  19. Kolmogorov, V.; Zabih, R. Multi-camera scene reconstruction via graph cuts. In Proceedings of the European Conference on Computer Vision, Copenhagen, Denmark, 28–31 May 2002; Springer: Berlin/Heidelberg, Germany, 2002; pp. 82–96. [Google Scholar]
  20. Woodford, O.; Torr, P.; Reid, I.; Fitzgibbon, A. Global stereo reconstruction under second-order smoothness priors. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 2115–2128. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Bleyer, M.; Rother, C.; Kohli, P. Surface stereo with soft segmentation. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 1570–1577. [Google Scholar]
  22. Johannsen, O.; Sulc, A.; Goldluecke, B. What Sparse Light Field Coding Reveals about Scene Structure. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3262–3270. [Google Scholar] [CrossRef]
  23. Wanner, S.; Goldluecke, B. Reconstructing reflective and transparent surfaces from epipolar plane images. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); LNCS: Tokyo, Japan, 2013; Volume 8142, pp. 1–10. [Google Scholar] [CrossRef] [Green Version]
  24. Criminisi, A.; Kang, S.B.; Swaminathan, R.; Szeliski, R.; Anandan, P. Extracting layers and analyzing their specular properties using epipolar-plane-image analysis. Comput. Vis. Image Underst. 2005, 97, 51–85. [Google Scholar] [CrossRef]
  25. Wanner, S.; Goldluecke, B. Globally Consistent Depth Labeling of 4D Light Fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 41–48. [Google Scholar] [CrossRef]
  26. Zhu, X.; Cohen, S.; Schiller, S.; Milanfar, P. Estimating Spatially Varying Defocus Blur from a Single Image. IEEE Trans. Image Process. 2013, 22, 4879–4891. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Zhuo, S.; Sim, T. Defocus map estimation from a single image. Pattern Recognit. 2011, 44, 1852–1858. [Google Scholar] [CrossRef]
  28. Nayar, S.K.; Nakagawa, Y. Shape from focus. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 824–831. [Google Scholar] [CrossRef] [Green Version]
  29. Tao, M.W.; Hadap, S.; Malik, J.; Ramamoorthi, R. Depth from Combining Defocus and Correspondence Using Light-Field Cameras. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 673–680. [Google Scholar] [CrossRef]
  30. Heber, S.; Pock, T. Convolutional Networks for Shape from Light Field. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3746–3754. [Google Scholar] [CrossRef]
  31. 3dMD Laser Scanner. 2018. Available online: https://3dmd.com (accessed on 18 December 2021).
  32. Honauer, K.; Johannsen, O.; Kondermann, D.; Goldluecke, B. A dataset and evaluation methodology for depth estimation on 4d light fields. In Proceedings of the Asian Conference on Computer Vision, Taipei, Taiwan, 20–24 November 2016; Springer: Berlin/Heidelberg, Germany, 2017; pp. 19–34. [Google Scholar]
  33. Rerabek, M.; Ebrahimi, T. New light field image dataset. In Proceedings of the 8th International Conference on Quality of Multimedia Experience (QoMEX), Lisbon, Portugal, 6–8 June 2016. [Google Scholar]
Figure 1. (a) The sub-aperture image view; (b) the EPI for the vertical line represented in (a); (c) the EPI for the horizontal line represented in (a).
Figure 1. (a) The sub-aperture image view; (b) the EPI for the vertical line represented in (a); (c) the EPI for the horizontal line represented in (a).
Sensors 22 01993 g001
Figure 2. Diagrammatic representation of a lenslet-based plenoptic camera to capture light field images [17].
Figure 2. Diagrammatic representation of a lenslet-based plenoptic camera to capture light field images [17].
Sensors 22 01993 g002
Figure 3. Schematic representation of the plenoptic function [18].
Figure 3. Schematic representation of the plenoptic function [18].
Sensors 22 01993 g003
Figure 4. The three representations of the two-plane light field. (a) The sub-aperture images or the pinhole view. (b) The sub-view for the light field. (c) Epipolar-plane image, which is obtained by fixing both spatial and angular co-ordinates [18].
Figure 4. The three representations of the two-plane light field. (a) The sub-aperture images or the pinhole view. (b) The sub-view for the light field. (c) Epipolar-plane image, which is obtained by fixing both spatial and angular co-ordinates [18].
Sensors 22 01993 g004
Figure 5. Flow of the proposed algorithm.
Figure 5. Flow of the proposed algorithm.
Sensors 22 01993 g005
Figure 6. Conceptual model for refocusing LF image at a different depth [17]. The (u,v) and (s,t) plane are surfaces of the camera, respectively, and (s′,t′) is the refocus plane.
Figure 6. Conceptual model for refocusing LF image at a different depth [17]. The (u,v) and (s,t) plane are surfaces of the camera, respectively, and (s′,t′) is the refocus plane.
Sensors 22 01993 g006
Figure 7. (a) Central sub-aperture image of a LF image, (b) focal stack image refocused on the background using median pixel values and (c) focal stack image refocused on the background using averaged pixel values and magnified region at depth discontinuity for each image.
Figure 7. (a) Central sub-aperture image of a LF image, (b) focal stack image refocused on the background using median pixel values and (c) focal stack image refocused on the background using averaged pixel values and magnified region at depth discontinuity for each image.
Sensors 22 01993 g007
Figure 8. (a) Central sub-aperture image of a LF image, (b) unrefined depth map calculated without gradient addition, (c) unrefined depth map calculated with gradient addition, (d) ground-truth depth map, (e) refined depth map calculated without gradient addition and (f) refined depth map calculated with gradient addition. The Badpix metric for without and with gradient addition is 9.11% and 96.23%, respectively.
Figure 8. (a) Central sub-aperture image of a LF image, (b) unrefined depth map calculated without gradient addition, (c) unrefined depth map calculated with gradient addition, (d) ground-truth depth map, (e) refined depth map calculated without gradient addition and (f) refined depth map calculated with gradient addition. The Badpix metric for without and with gradient addition is 9.11% and 96.23%, respectively.
Sensors 22 01993 g008
Figure 9. The red and green squares are the two overlap** 4 × 4 pixel patches used to cover the entire image. As the patches overlap, only the highlighted red and the green pixels from the red and green squares are used to estimate the depth.
Figure 9. The red and green squares are the two overlap** 4 × 4 pixel patches used to cover the entire image. As the patches overlap, only the highlighted red and the green pixels from the red and green squares are used to estimate the depth.
Sensors 22 01993 g009
Figure 10. (a) Central sub-aperture image of a LF image, (b) a magnified 6 × 6 RGB image patch and (c) FFT of the image patch.
Figure 10. (a) Central sub-aperture image of a LF image, (b) a magnified 6 × 6 RGB image patch and (c) FFT of the image patch.
Sensors 22 01993 g010
Figure 11. (a) The RGB image patch and (b) the FFT patch at different focal lengths. The patch with the red boundary is the closest match to the reference patch in Figure 10. (c) The graph shows the MSE values for the central image in Figure 10, with the corresponding focal stack image patch.
Figure 11. (a) The RGB image patch and (b) the FFT patch at different focal lengths. The patch with the red boundary is the closest match to the reference patch in Figure 10. (c) The graph shows the MSE values for the central image in Figure 10, with the corresponding focal stack image patch.
Sensors 22 01993 g011
Figure 12. Comparison between the ground truth and the estimated depth map before and after the refinement step for synthetic images.
Figure 12. Comparison between the ground truth and the estimated depth map before and after the refinement step for synthetic images.
Sensors 22 01993 g012
Figure 13. Comparison between the estimated depth map before and after the refinement step for real images.
Figure 13. Comparison between the estimated depth map before and after the refinement step for real images.
Sensors 22 01993 g013
Figure 14. Visual comparison between ground truth depth map, the result using FFT to estimate depth map (proposed algorithm) and depth map estimated using RGB patches.
Figure 14. Visual comparison between ground truth depth map, the result using FFT to estimate depth map (proposed algorithm) and depth map estimated using RGB patches.
Sensors 22 01993 g014
Figure 15. Visual comparison for the dot image of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for synthetic LF images.
Figure 15. Visual comparison for the dot image of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for synthetic LF images.
Sensors 22 01993 g015
Figure 16. Misdetection of textureless regions in Rosemary image.
Figure 16. Misdetection of textureless regions in Rosemary image.
Sensors 22 01993 g016
Figure 17. MSE for the central image patch and the patch at different focal lengths when the depth is estimated correctly.
Figure 17. MSE for the central image patch and the patch at different focal lengths when the depth is estimated correctly.
Sensors 22 01993 g017
Figure 18. (a) The depth map of cotton image with the red square showing the cross error patch. (b) The central image with the red box showing the error patch. (c) The MSE of the central image patch with the patch at different focal length.
Figure 18. (a) The depth map of cotton image with the red square showing the cross error patch. (b) The central image with the red box showing the error patch. (c) The MSE of the central image patch with the patch at different focal length.
Sensors 22 01993 g018
Figure 19. Visual comparison of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for synthetic LF images.
Figure 19. Visual comparison of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for synthetic LF images.
Sensors 22 01993 g019aSensors 22 01993 g019b
Figure 20. Visual comparison of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for synthetic LF images.
Figure 20. Visual comparison of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for synthetic LF images.
Sensors 22 01993 g020aSensors 22 01993 g020b
Figure 21. Visual comparison of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for real LF images.
Figure 21. Visual comparison of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for real LF images.
Sensors 22 01993 g021aSensors 22 01993 g021b
Figure 22. Visual comparison of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for real LF images.
Figure 22. Visual comparison of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for real LF images.
Sensors 22 01993 g022aSensors 22 01993 g022b
Figure 23. Visual comparison of the ground truth with the proposed algorithm, Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for dots image for Badpix 0.03 and 0.01.
Figure 23. Visual comparison of the ground truth with the proposed algorithm, Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for dots image for Badpix 0.03 and 0.01.
Sensors 22 01993 g023aSensors 22 01993 g023b
Figure 24. Visual comparison of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for synthetic LF images with added noise.
Figure 24. Visual comparison of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for synthetic LF images with added noise.
Sensors 22 01993 g024aSensors 22 01993 g024b
Figure 25. Visual comparison of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for synthetic LF images with added noise.
Figure 25. Visual comparison of the proposed algorithm with Strecke et al. [9], Wang et al. [10], Zhang et al. [11] and Shin et al. [12] for synthetic LF images with added noise.
Sensors 22 01993 g025aSensors 22 01993 g025b
Table 1. Comparison of the estimated depth map when using the FFT Patch and RGB Patch algorithms.
Table 1. Comparison of the estimated depth map when using the FFT Patch and RGB Patch algorithms.
  Dots  Dots
FFT Patch Depth MapRGB Patch Depth Map
  Badpix 0.070.97050.7760
  Badpix 0.030.88530.5967
  Badpix 0.010.38800.2391
Table 2. Quantitative depth map comparison for synthetic data to ground truth for different algorithms.
Table 2. Quantitative depth map comparison for synthetic data to ground truth for different algorithms.
Back-GammonDotsKitchenMedi-eval2Museum
Proposed Results
Badpix70.82300.96050.70100.93620.8440
Badpix30.73240.88530.59410.85280.7772
Badpix10.49100.38800.37490.55140.5305
Strecke et al. [9]
Badpix70.95800.62730.72240.96080.8578
Badpix30.92830.45140.62820.88950.7615
Badpix10.66060.17770.46440.64690.5256
Wang et al. [10]
Badpix70.87530.88010.63000.51360.8522
Badpix30.45250.24850.39910.11190.6902
Badpix10.05440.04560.17720.03700.2741
Zhang et al. [11]
Badpix70.78890.73580.63790.95800.8940
Badpix30.37620.48100.31650.75130.5413
Badpix10.10570.48100.09970.26580.1899
Shin et al. [12]
Badpix70.97770.94730.79310.98470.9598
Badpix30.95940.79570.72090.95840.9053
Badpix10.82650.51220.48090.72630.6478
Table 3. Quantitative depth map comparison for synthetic data to ground truth for different algorithms.
Table 3. Quantitative depth map comparison for synthetic data to ground truth for different algorithms.
PillowsPlatonicPyramidsStripesTomb
Proposed Results
Badpix70.92120.97470.99200.88530.9696
Badpix30.87690.94470.95820.82750.9100
Badpix10.60960.76000.74850.67320.6423
Strecke et al. [9]
Badpix70.97100.96450.99690.87410.9813
Badpix30.86870.92300.99270.85560.9252
Badpix10.49140.77920.94170.49250.6875
Wang et al. [10]
Badpix70.93870.65830.98430.82310.7953
Badpix30.56110.46200.75200.00480.4134
Badpix10.14920.18890.07370.00040.1359
Zhang et al. [11]
Badpix70.93980.99060.89580.83730.9622
Badpix30.50660.74540.18850.52430.7500
Badpix10.18690.29460.06340.52430.2871
Shin et al. [12]
Badpix70.99390.99810.99720.98940.9963
Badpix30.97720.99410.99170.98650.9826
Badpix10.77270.72730.86730.88690.6453
Table 4. Comparison for the ‘dot’ image with ground truth for proposed and state-of-the-art algorithms.
Table 4. Comparison for the ‘dot’ image with ground truth for proposed and state-of-the-art algorithms.
Dots
Proposed ResultStrecke et al. [9]Wang et al. [10]Zhang et al. [11]Shin et al. [12]
Badpix70.96050.62730.88000.73570.9473
Badpix30.88530.45140.24850.48100.7957
Badpix10.38800.17770.04560.48100.5122
Table 5. Quantitative depth map comparison for synthetic data to ground truth for different algorithms for noisy images.
Table 5. Quantitative depth map comparison for synthetic data to ground truth for different algorithms for noisy images.
Back-GammonDotsKitchenMedi-eval2Museum
Proposed Results
Badpix70.74080.96200.63410.81710.6921
Badpix30.51260.85610.43230.55840.4889
Badpix10.21010.28080.17330.22900.2023
Strecke et al. [9]
Badpix70.27810.39750.23090.30240.1938
Badpix30.13120.18950.11400.14190.0919
Badpix10.04500.06340.04020.04870.0318
Wang et al. [10]
Badpix70.00220.86190.12290.08920.1868
Badpix30.00080.76840.05700.03400.0790
Badpix10.00030.13510.01970.00830.0260
Zhang et al. [11]
Badpix70.01440.00020.15870.24560.1054
Badpix30.00570.00010.06660.10220.0517
Badpix10.00190.00000.02170.03220.0174
Shin et al. [12]
Badpix70.57780.89900.50350.65120.5237
Badpix30.32650.66240.30900.38980.3112
Badpix10.11620.30340.12470.14510.1181
Table 6. Quantitative depth map comparison for synthetic data to ground truth for different algorithms for noisy images.
Table 6. Quantitative depth map comparison for synthetic data to ground truth for different algorithms for noisy images.
PillowsPlatonicPyramidsStripesTomb
Proposed Results
Badpix70.68230.84570.98910.35820.8008
Badpix30.50990.62720.91080.19820.5089
Badpix10.24170.26000.48350.09300.1877
Strecke et al. [9]
Badpix70.32120.30930.65570.13880.1404
Badpix30.16980.14760.34560.06410.0615
Badpix10.06130.05110.11360.02120.0206
Wang et al. [10]
Badpix70.14720.25150.61360.24880.0561
Badpix30.07000.11260.08700.11060.0209
Badpix10.02570.03660.01410.00140.0069
Zhang et al. [11]
Badpix70.17940.03940.25690.04190.0530
Badpix30.08140.01480.09170.02190.0214
Badpix10.02730.00450.01150.00780.0075
Shin et al. [12]
Badpix70.60000.66210.97290.20840.3957
Badpix30.43830.38360.83250.10510.1904
Badpix10.21930.14610.41320.03720.0664
Table 7. Average runtime for synthetic data.
Table 7. Average runtime for synthetic data.
Synthetic Light Field Dataset
ProposedStrecke et al. [9]Wang et al. [10]Zhang et al. [11]Shin et al. [12]
1418 s462 s243 s537 s7 s
Table 8. Average runtime for real data.
Table 8. Average runtime for real data.
Real Light Field Dataset
ProposedStrecke et al. [9]Wang et al. [10]Zhang et al. [11]Shin et al. [12]
2559 s187 s658 s1413 s8.43 s
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sharma, R.; Perry, S.; Cheng, E. Noise-Resilient Depth Estimation for Light Field Images Using Focal Stack and FFT Analysis. Sensors 2022, 22, 1993. https://doi.org/10.3390/s22051993

AMA Style

Sharma R, Perry S, Cheng E. Noise-Resilient Depth Estimation for Light Field Images Using Focal Stack and FFT Analysis. Sensors. 2022; 22(5):1993. https://doi.org/10.3390/s22051993

Chicago/Turabian Style

Sharma, Rishabh, Stuart Perry, and Eva Cheng. 2022. "Noise-Resilient Depth Estimation for Light Field Images Using Focal Stack and FFT Analysis" Sensors 22, no. 5: 1993. https://doi.org/10.3390/s22051993

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop