Depth extraction of 3D objects using axially

Depth extraction of 3D objects using axially
distributed image sensing
Suk-Pyo Hong,1Donghak Shin,2,*Byung-Gook Lee,2 and Eun-Soo Kim1
1
2
HoloDigilog Human Media Research Center(HoloDigilog), 3D Display Research Center(3DRC), Kwangwoon
University, Wolgye-Dong, Nowon-Gu, Seoul 139-701, South Korea
Institute of Ambient Intelligence, Dongseo University, 47, Jurye-Ro, Sasang-Gu, Busan 617-716, South Korea
*
[email protected]
Abstract: Axially distributed image sensing (ADS) technique is capable of
capturing 3D objects and reconstructing high-resolution slice plane images
for 3D objects. In this paper, we propose a computational method for depth
extraction of 3D objects using ADS. In the proposed method, the highresolution elemental images are recorded by simply moving the camera
along the optical axis and the recorded elemental images are used to
generate a set of 3D slice images using the computational reconstruction
algorithm based on ray back-projection. To extract depth of 3D object, we
propose the simple block comparison algorithm between the first elemental
image and a set of 3D slice images. This provides a simple computation
process and robustness for depth extraction. To demonstrate our method,
we carry out the preliminary experiments of three scenarios for 3D objects
and the results are presented. To our best knowledge, this is the first report
to extract the depth information using an ADS method.
©2012 Optical Society of America
OCIS codes: (100.6890) Three-dimensional image processing; (110.6880) Three-dimensional
image acquisition; (110.4190) Multiple imaging.
References and links
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
T. Okoshi, Three-dimensional Imaging Techniques (Academic Press, 1976).
J.-S. Ku, K.-M. Lee, and S.-U. Lee, “Multi-image matching for a general motion stereo camera model,” Pattern
Recognit. 34(9), 1701–1712 (2001).
A. Stern and B. Javidi, “Three-dimensional image sensing, visualization, and processing using integral imaging,”
Proc. IEEE 94(3), 591–607 (2006).
K. Itoh, W. Watanabe, H. Arimoto, and K. Isobe, “Coherence-based 3-D and spectral imaging and laserscanning microscopy,” Proc. IEEE 94(3), 608–628 (2006).
J.-H. Park, K. Hong, and B. Lee, “Recent progress in three-dimensional information processing based on integral
imaging,” Appl. Opt. 48(34), H77–H94 (2009).
Y.-M. Kim, K.-H. Hong, and B. Lee, “Recent researches based on integral imaging display method,” 3D
Research 1(1), 17–27 (2010).
M. DaneshPanah, B. Javidi, and E. A. Watson, “Three dimensional imaging with randomly distributed sensors,”
Opt. Express 16(9), 6368–6377 (2008).
D. Shin and B. Javidi, “3D visualization of partially occluded objects using axially distributed sensing,” J. Disp.
Technol. 7(5), 223–225 (2011).
D. Shin and B. Javidi, “Three-dimensional imaging and visualization of partially occluded objects using axially
distributed stereo image sensing,” Opt. Lett. 37(9), 1394–1396 (2012).
J.-H. Park, S. Jung, H. Choi, Y. Kim, and B. Lee, “Depth extraction by use of a rectangular lens array and onedimensional elemental image modification,” Appl. Opt. 43(25), 4882–4895 (2004).
J.-H. Park, Y. Kim, J. Kim, S.-W. Min, and B. Lee, “Three-dimensional display scheme based on integral
imaging with three-dimensional information processing,” Opt. Express 12(24), 6020–6032 (2004).
G. Passalis, N. Sgouros, S. Athineos, and T. Theoharis, “Enhanced reconstruction of three-dimensional shape
and texture from integral photography images,” Appl. Opt. 46(22), 5311–5320 (2007).
D.-H. Shin, B.-G. Lee, and J.-J. Lee, “Occlusion removal method of partially occluded 3D object using subimage block matching in computational integral imaging,” Opt. Express 16(21), 16294–16304 (2008).
J.-H. Jung, K. Hong, G. Park, I. Chung, J.-H. Park, and B. Lee, “Reconstruction of three-dimensional occluded
object using optical flow and triangular mesh reconstruction in integral imaging,” Opt. Express 18(25), 26373–
26387 (2010).
#172985 - $15.00 USD
(C) 2012 OSA
Received 20 Jul 2012; revised 17 Sep 2012; accepted 18 Sep 2012; published 24 Sep 2012
8 October 2012 / Vol. 20, No. 21 / OPTICS EXPRESS 23044
15. D.-C. Hwang, D.-H. Shin, S.-C. Kim, and E.-S. Kim, “Depth extraction of three-dimensional objects in space by
the computational integral imaging reconstruction technique,” Appl. Opt. 47(19), D128–D135 (2008).
16. B.-G. Lee, H.-H. Kang, and E.-S. Kim, “Occlusion removal method of partially occluded object using variance
in computational integral imaging,” 3D Research 1(2), 2–10 (2010).
17. M. DaneshPanah and B. Javidi, “Profilometry and optical slicing by passive three-dimensional imaging,” Opt.
Lett. 34(7), 1105–1107 (2009).
18. J.-S. Jang and B. Javidi, “Three-dimensional synthetic aperture integral imaging,” Opt. Lett. 27(13), 1144–1146
(2002).
19. A. Stern and B. Javidi, “3-D computational synthetic aperture integral imaging (COMPSAII),” Opt. Express
11(19), 2446–2451 (2003).
20. Y. S. Hwang, S.-H. Hong, and B. Javidi, “Free view 3-D visualization of occluded objects by using
computational synthetic aperture integral imaging,” J. Disp. Technol. 3(1), 64–70 (2007).
21. M. Cho and B. Javidi, “Three-dimensional visualization of objects in turbid water using integral imaging,” J.
Disp. Technol. 6(10), 544–547 (2010).
22. R. Schulein, M. DaneshPanah, and B. Javidi, “3D imaging with axially distributed sensing,” Opt. Lett. 34(13),
2012–2014 (2009).
23. D. Shin, M. Cho, and B. Javidi, “Three-dimensional optical microscopy using axially distributed image sensing,”
Opt. Lett. 35(21), 3646–3648 (2010).
24. D. Shin and B. Javidi, “Visualization of 3D objects in scattering medium using axially distributed sensing,” J.
Disp. Technol. 8(6), 317–320 (2012).
1. Introduction
Acquiring depth information from three-dimensional (3D) objects in real world is an
important issue in many diverse fields of computer vision, 3D display and 3D recognition.
Three-dimensional (3D) passive imaging makes it possible to extract depth information by
recording different perspectives of 3D objects [1–9]. Among 3D passive imaging
technologies, integral imaging has been widely studied for depth extraction [10–17]. In
original structure of integral imaging, the lens array is used to record 3D objects. However,
the recorded elemental images have low-resolution so that the extracted depth is poor. To
overcome this problem, synthetic aperture integral imaging (SAII) was proposed [18–21], in
which the multiple cameras or moving cameras are used to acquire depth information. Even
though it can provide high-resolution depth information, the system structure of integral
imaging is complex because of 2D grid array structure of many cameras or moving a camera
along both horizontal and vertical direction. Recently, a simple modification structure of SAII
was proposed. This is called an axially distributed image sensing (ADS) method [8,22–24], in
which longitudinal perspective information is obtained by translating a camera along its
optical axis. The high-resolution elemental images recorded from ADS can be used for
acquiring high-resolution depth information of real-world 3D objects. However, it has a
disadvantage of the limitation of the object pickup zone. Despite this limitation, the ADS is
an attractive way to provide high-resolution elemental images and simple axial movement.
In this paper, we propose a new depth extraction method of 3D objects using ADS. In the
proposed method, we first record the high-resolution elemental images by moving the camera
along the optical axis. Then, the recorded elemental images are reconstructed as a set of slice
plane images using the computational reconstruction algorithm based on ray back-projection.
Finally the depth information of 3D objects is extracted by the block comparison algorithm
between first elemental image and many slice plane images. The proposed depth extraction is
implemented by simple summation of elemental images and comparison of intensity
distribution between blocks from first elemental image and slice plane images. This can avoid
the complex corresponding search process in the conventional depth extraction algorithm.
Therefore our method provides simple computation process and robustness for depth
extraction. To show the usefulness of the proposed method, we carry out the preliminary
experiments of 3D objects and the results are presented.
2. Depth extraction method using ADS
The proposed depth extraction method using ADS is shown in Fig. 1. It is divided into three
sub-parts: ADS pickup part, digital reconstruction part and the depth extraction part.
#172985 - $15.00 USD
(C) 2012 OSA
Received 20 Jul 2012; revised 17 Sep 2012; accepted 18 Sep 2012; published 24 Sep 2012
8 October 2012 / Vol. 20, No. 21 / OPTICS EXPRESS 23045
Fig. 1. Proposed depth extraction processes using ADS
2.1 ADS pickup
The ADS pickup part of 3D objects is shown in Fig. 2, in which a single camera records
elemental images by moving the camera along the optical axis [7].
Fig. 2. Pickup process in ADS
We suppose that the focal length of imaging lens is defined as g. Then, the different
elemental images are captured along optical axis (z axis) if 3D objects are located at a
distance z1 away from the first camera. We can record k elemental images by shifting a
camera with k-1 times. Then, the last camera is located at zk. If Δz is the separation distance
between two adjacent cameras in ADS pickup process as shown in Fig. 2, the ith elemental
image is captured from the camera located at a distance of di = Z-z1-(i-1)Δz from the object.
Since we can capture each elemental image at a different camera position, it contains the
object image with different scale level. In other words, when i = 1, the object image is the
smallest in elemental image because the camera position is farthest from 3D object. While,
the object image is largest when i = k.
2.2 Slice image reconstruction using the recorded elemental images
The second part of the proposed depth extraction method is digital reconstruction. The aim of
the digital reconstruction part is to generate slice plane image using the recorded elemental
images from the first part of ADS pickup. The digital reconstruction process is shown in Fig.
3. The digital reconstruction process of 3D objects is the inverse process of ADS pickup. It
can be implemented on the basis of an inverse mapping procedure through a pinhole model
[7]. Each camera in the ADS pickup part is modeled as pinhole and elemental image at the
original camera position as shown in Fig. 3. We assume that a reconstruction plane for the
computational reconstruction of the 3D image is located at distance z = L. Each elemental
image is inversely projected through each corresponding pinhole to the reconstruction plane
at L. Then, the ith inversely projected elemental image is magnified with mi = (L-zi)/g. At the
reconstruction plane, all inversely mapped elemental images are superimposed each other
with the different magnifications. In Fig. 3, we assume that Ei is the ith elemental image with
the size of p × q, and IL is the superimposed image of all the inversely mapped image of the
elemental image at the reconstruction plane L. IL is given by
#172985 - $15.00 USD
(C) 2012 OSA
Received 20 Jul 2012; revised 17 Sep 2012; accepted 18 Sep 2012; published 24 Sep 2012
8 October 2012 / Vol. 20, No. 21 / OPTICS EXPRESS 23046
IL =
1
k
k
U
i =1
i
Ei
(1)
where Ui is upsampling operator for magnification of Ei at the reconstruction plane of z = L-zi
and the size of IL is m1p × m1q.
Since the superimposition of all elemental images can require large computation load due
to the large magnification factor, we can modify Eq. (1) by using the downsampling operator
Dr of image by a factor of r. Then, it becomes
IL ≈
1
k
k
D U
i =1
r
i
Ei
(2)
In order to generate the 3D volume information, we repeat this process for different distances.
Fig. 3. Principle of digital reconstruction in ADS
2.3 Depth extraction process
In the depth extraction part of the proposed method, we extract depth information using many
slice plane images from digital reconstruction described in Eq. (2). The reconstructed slice
plane images consist of different mixture images with focused images and blurred images
according to the reconstruction distances. The focused images of 3D object are reconstructed
only at the original position of 3D object. While, blurred images are shown out of original
position. Based on this principle, we want to find the focused image part in the reconstructed
slice plane images.
The depth extraction algorithm used in this paper is shown in Fig. 4. This is based on
separation between focused image and defocused image. To find the focused image part in
each slice plane image, we use block comparison between the first elemental image and slice
plane images. Note that the first elemental image is composed of only focused images of 3D
objects. We select block images in both first elemental image and the slice plane images. The
block images have the same position as shown in Fig. 4. Then we apply block comparison
algorithm into two selective block images. Here, block comparison calculates intensity error
between the block image of the first elemental image (DrU1E1) and a set of block images of
slice images (IL) reconstructed at L. The intensity error between the blocks along the
reconstruction distance (L) can be defined as the sum of absolute difference (SAD):
#172985 - $15.00 USD
(C) 2012 OSA
Received 20 Jul 2012; revised 17 Sep 2012; accepted 18 Sep 2012; published 24 Sep 2012
8 October 2012 / Vol. 20, No. 21 / OPTICS EXPRESS 23047
b
b
SAD L =  | D rU 1E 1 (x + i , y + j ) − I L (x + i , y + j ) |
(3)
i =1 j =1
where the block size is b × b.
Using SAD result, the depth of each point (x,y) can be extracted by finding L value where
SADL is minimized over the range of L. This can be mathematically formulated as:
Lˆ (x , y ) = arg min SAD L (x , y )
(4)
L
where the size of Lˆ is m1p/r × m1q/r.
The depth map can be obtained by calculating depth of all points in the images.
Fig. 4. Block comparison algorithm between the first elemental image and slice images
2.4 Depth accuracy in ADS
The accuracy of extracted depth information is dependent of the step size of L value. We can
calculate the minimum step size of δ value using Fig. 5. As shown in Fig. 5, when we suppose
two L values (L and L + δ), the inversely magnified elemental image is superimposed at the
reconstruction plane. If the object is located a distance d from the axis and c is the pixel size
of the camera sensor, the elemental image should be superimposed at different pixel of the
reconstructed plane to separate the reconstructed plane image. The minimum step size of δ
value is obtained when the pixel position in the reconstruction plane has the difference of 1
pixel. Then, it becomes
δ=
(L − z 1 )c
.
d
(5)
For practical parameters, we simulated Eq. (5) as shown in Fig. 6. As d increases, δ can be
reduced. When c = 5 μm, d = 100 mm, L = 500 mm, and z1 = 300 mm, the minimum step size
δ becomes 10 μm.
#172985 - $15.00 USD
(C) 2012 OSA
Received 20 Jul 2012; revised 17 Sep 2012; accepted 18 Sep 2012; published 24 Sep 2012
8 October 2012 / Vol. 20, No. 21 / OPTICS EXPRESS 23048
Fig. 5. Ray diagram for calculation of minimum step size
Fig. 6. Calculation results of minimum step size for practical parameters.
3. Experiments and results
To demonstrate our depth extraction method using ADS, we performed the preliminary
experiments for three scenarios composed of different 3D objects. The experimental structure
is shown in Fig. 7.
#172985 - $15.00 USD
(C) 2012 OSA
Received 20 Jul 2012; revised 17 Sep 2012; accepted 18 Sep 2012; published 24 Sep 2012
8 October 2012 / Vol. 20, No. 21 / OPTICS EXPRESS 23049
Fig. 7. Three scenarios for optical experiments object
First objects were composed of simple characters ‘k’ and ‘w’. Second scenario has two
objects of ‘car’ and ‘sign’ toys. Last one is three ‘car’ objects with continuous depth. Front
object and behind object are positioned between approximately 350 mm and 500 away from
the first camera, respectively. They are located at approximately 70 mm from the optical axis.
We use a Nikon camera (D7000) whose pixels are 3872 × 2592. The imaging lens with focal
length f = 40 mm is used in this experiments. The camera is translated at Δz = 3 mm
increments for a total of K = 41 elemental images and a total displacement distance of 120
mm. Among them, some elemental images for second scenario are shown in Fig. 8.
Fig. 8. Four examples of the recorded elemental images for second object (a) first elemental
image (b) 10th elemental image (c) 20th elemental image (d) 30th elemental image
After recording the elemental images using ADS, we reconstruct slice plane image for 3D
objects. The recorded 41 elemental images are used to the digital reconstruction algorithm
using Eq. (2). In digital reconstruction process, the pinhole gap was 40 mm and the interval of
reconstruction plane was 10 mm, which is much larger than the minimum step size of Eq. (5).
To simplify the computational calculations, the downsampling factor was 1/U1 and then Ui
was normalized by U1. The reconstruction plane was moved from 300 mm to 600 mm. Then
we can reconstruct 31 slice plane images. Some reconstructed slice plane images for second
objects of Fig. 7(b) are shown in Fig. 9.
#172985 - $15.00 USD
(C) 2012 OSA
Received 20 Jul 2012; revised 17 Sep 2012; accepted 18 Sep 2012; published 24 Sep 2012
8 October 2012 / Vol. 20, No. 21 / OPTICS EXPRESS 23050
Fig. 9. Reconstructed slice images for second objects (a) at 350 mm (b) at 420 mm (c) at 500
mm.
The ‘car’ object was focused at L = 350 as shown in Fig. 9(a).While, the ‘sign’ object was
generated clearly at L = 500 mm as shown in Fig. 9(c). And, the blurred image was obtained
at L = 420 mm as shown in Fig. 9(b).
Now, we estimated the depth of the object using the first reference elemental image and
31 slice plane images. We apply the block comparison algorithm to these images. The block
size was 8 × 8. The estimated depths are shown in Fig. 10. The depth results were well
extracted. This result reveals that the proposed method can extract the 3D information of
object effectively.
Fig. 10. Extracted depth map for three scenarios
4. Conclusion
In conclusion, we have presented a depth extraction method using ADS. In the proposed
method, the high-resolution elemental images were recorded by simply moving the camera
along the optical axis and the recorded elemental images were used to generate a set of 3D
slice images using the computational reconstruction algorithm based on ray back-projection.
To extract depth of 3D object, the block comparison algorithm between the first elemental
image and a set of 3D slice images was used. Since ADS provides high-resolution slice plane
images of 3D objects, we can extract high-resolution 3D depth information. We performed
the preliminary experiment of 3D object and presented the results to show the usefulness of
the proposed method.
#172985 - $15.00 USD
(C) 2012 OSA
Received 20 Jul 2012; revised 17 Sep 2012; accepted 18 Sep 2012; published 24 Sep 2012
8 October 2012 / Vol. 20, No. 21 / OPTICS EXPRESS 23051
Acknowledgment
This work was supported by the National Research Foundation of Korea (NRF) grant funded
by the Korea government (MEST) (No. 2012-0009223)
#172985 - $15.00 USD
(C) 2012 OSA
Received 20 Jul 2012; revised 17 Sep 2012; accepted 18 Sep 2012; published 24 Sep 2012
8 October 2012 / Vol. 20, No. 21 / OPTICS EXPRESS 23052