Depth Resolution Analysis of Axially Distributed Stereo Camera Systems under Fixed Constrained Resources
 Author: Cho Myungjin, Shin Donghak
 Organization: Cho Myungjin; Shin Donghak
 Publish: Current Optics and Photonics Volume 17, Issue6, p500~505, 25 Dec 2013

ABSTRACT
In this paper, we propose a novel framework to evaluate the depth resolution of axially distributed stereo sensing (ADSS) under fixed resource constraints. The proposed framework can evaluate the performance of ADSS systems based on various sensing parameters such as the number of cameras, the number of total pixels, pixel size and so on. The Monte Carlo simulations for the proposed framework are performed and the evaluation results are presented.

KEYWORD
3D imaging , Axially distributed stereo sensing , Depth resolution , Elemental images

I. INTRODUCTION
Threedimensional (3D) imaging techniques have been considered as an important issue in the computer vision, target tracking, object recognition, and so on [15].Various methods for capturing and visualization of 3D objects in space have been studied [614] including integral imaging, synthetic aperture integral imaging (SAII) and axially distributed image sensing (ADS). Among them, the extended version of ADS was reported called axially distributed stereo sensing (ADSS) which is implemented using a stereo camera [14]. In this method, a stereo camera is translated along its optical axis to obtain multiple image pairs and the computational reconstruction is implemented by the uniform superposition of the resized elemental image pairs. This can solve the problem that the collection of 3D information is not uniform across the sensor in the conventional ADS system.
Recently, the resolution analysis methods for various 3D imaging systems under the equallyconstrained resources have been reported [1518]. In 2012, the
N ocular imaging system was firstly analyzed using two point source resolution criteria. Here, the lateral and longitudinal resolutions of theN ocular imaging systems were analyzed based on several factors such as the number of sensors, pixel size, imaging optics, relative sensor configuration, and so on [15]. In 2013, we proposed the resolution analysis of the ADS method based on two point source resolution criteria [18]. The analysis results for system parameters were presented.In this paper, we propose a new framework for performance evaluation of ADSS systems under the equallyconstrained resources. The evaluation parameters are a fixed number of pixels, a fixed moving distance, a fixed pixel size, and so on. For the resolution analysis of the proposed ADSS framework, we use the two point sources resolution criterion [15]. We evaluate the depth resolution through the Monte Carlo simulations.
II. REVIEW OF ADSS
In general, the ADSS method is composed of a pickup part and a digital reconstruction part [14]. Figure 1 shows the total system structure of the ADSS. In the pickup part of the ADSS as shown in Fig. 1, we record an elemental image pair using stereo cameras with different distances. And, the distance between the two cameras is
b . Then, we capture multiple elemental image pairs by moving the stereo camera along its optical axis.With the recorded multiple elemental image pairs, we can generate 3D sliced images in the digital reconstruction part of the ADSS which is shown in Fig. 2. The digital reconstruction process of 3D objects is the same with the inverse process of ADSS pickup. It can be implemented on the basis of an inverse mapping procedure through a virtual pinhole model as shown in Fig. 2.
Let us assume that the first pinhole is located at
z =0 and thek th pinhole is located atz =(k 1)Δz in Fig. 2. We define that and are the left and right images of thek th elemental image pair, respectively. Then,R_{k} (x ,y ,z_{r} ) is the inversely mapped image of thek th elemental image pair through thek th pinhole at the reconstruction image planez_{r} and it becomeswhere
M_{k} =[z_{r}  (k  1)Δz ] /g . Equation (1) means the uniform superposition of two resized elemental images. The final reconstructed 3D image at distancez_{r} is obtained by summation of all resized elemental images. That is,In fact, the conventional ADS system cannot capture the correct 3D information near the optical axis. However, the ADSS system can overcome this problem by generating 3D images near the optical axis.
III. RESOLUTION ANALYSIS FOR ADSS SYSTEM
Figure 3 shows the general framework of ADSS system with stereo camera located at
K different positions. In this framework, we want to use the equallyconstrained resources regardless of the number of stereo camera. To do so, let us assume that the total number of pixels (S ), the size of the pixel (c ), and the moving range (D ) are fixed. Thus, if we can use totalS pixels in the image sensors, the pixel number of each stereo camera becomesS/K . Also, we assume that the used imaging lenses are identical with the same focal length (f ) and the diameter (A ).In the ADSS framework as shown in Fig. 3, we can vary the positions of the stereo camera by
K different positions. The moving range is limited withinD . WhenK =1, the conventional stereo system can be constructed where two cameras with high resolution ofS /2 are used. WhenK >>1, the ADSS system can be constructed where multiple cameras structure with each camera having the low resolution ofS /2K is implemented.In this paper, we want to analyze the resolution performance for the ADSS framework as shown in Fig. 4. To do so, the resolution analysis method based on two point sources resolution criteria is utilized [15]. We modified this analysis method into stereo camera case in this paper.
For simplicity, we present onedimensional notation. We assume that there are two close point sources in space as shown in Fig. 4. We first find the mapping pixel index for the first point source located at (
x_{1} ,z_{1} ) as shown in Fig. 4. Then, the point spread function (PSF) of the imaging lens for the first point source is recorded into stereo camera (left and right cameras) located at the differenti th positions. We can calculate the center position of the PSF of the imaging lens for the first point source in thei th stereo camera. This becomeswhere
f is the focal length of the imaging lens and and are the positions of the left and right cameras ati th position. For the recorded point sources into and pixels, their pixel indexes are calculated bywhere⌈.⌉is the rounding operator.
Next, we calculate the unresolvable pixel area for the mapping pixels. The unresolvable pixel area means that two point sources cannot be separated within a single pixel. When the position of the second point source is located close to the first point source, we can resolve two point sources if the second point source is registered by at least one sensor so that it is on a pixel that is adjacent to the pixel that registered the first PSF. But, if the second PSF falls on the same pixel that recorded the PSF of the imaging lens for the first point source, we cannot resolve them. Based on this unresolved area, we can calculate the depth resolution.
Now let us consider the second point source located at (
x_{1} ,z_{2} ). Then, the possible mapping pixel ranges of the second point source for left and right cameras are given bywhere
δ is 1.22λf/A .Then, we want to find the unresolvable pixel area for the left and right cameras. The unresolvable pixel area of the second point source for the mapping pixel is calculated by using ray backprojection into the plane of two point sources as shown in Fig. 5. They are given by the following equations, respectively.
Finally, to calculate the depth resolution, we will find the common area for the unresolvable pixel areas calculated from all stereo cameras. The depth resolution can be considered as the common intersection of all unresolvable pixel ranges. Thus, the depth resolution for ADSS system becomes
IV. MONTE CARLO SIMULATIONS AND RESULTS
For the proposed framework of the ADSS system, the Monte Carlo simulations were performed by computer. We calculated the depth resolution statistically using the two point sources resolution analysis method. In our Monte Carlo simulation, two point sources were located longitudinally far from the stereo camera as shown in Fig. 4. We selected the random position of the first point source. And the second point source is moved in the longitudinal direction. Table 1 shows our experimental conditions of the Monte Carlo simulation for the proposed ADSS system.
We selected the random locations of two point sources with 4,000 trials for the Monte Carlo simulations. The depth resolutions for this experiment were calculated as the sample mean. For the several parameters, we carried out the simulations in terms of the depth resolution for ADSS frameworks. First experiment is the calculation of depth resolution according to the number of cameras. The result is shown in Fig. 6. According to the distance (
x ) between the optical axis and the point source plane, the results are plotted. From these results, we can observe that the depth resolution is improved asx becomes larger. Then, we calculated the depth resolution whenx =0 (optical axis case). This means that the depth resolution exists along the optical axis in the ADSS method because 3D perspective information can be recorded into either left or right camera. However, the depth resolution cannot exist in the conventional ADS system. As shown in Fig. 6, the depth resolution is improved as the number of stereo cameras increases because the use of many cameras is more effective for resolving two point sources in space. This means that the usage of more lowresolution stereo cameras may be useful to capture 3D objects in terms of the depth resolution compared with a single highresolution stereo camera.Figure 7 shows the depth resolution results according to the number of cameras and focal length of imaging lens when
x =0 mm andx =100 mm. The depth resolution is improved when a large number of cameras and large focal length are used. And, the calculation results of the depth resolution are shown in Figs. 810 when the camera pixel size, total number of camera pixels and moving range of stereo cameras are varied. The improved depth resolution is obtained when small pixel size and large moving range are used as shown in Fig. 8 and 9, in which we can see the large variations of depth resolution whenN =2. This is because the depth resolution was calculated based on twopixel information from a single stereo camera. This variation can be improved by using more position ranges of two point sources. However, it is seen that the total number of sensor pixels is not related to the depth resolution as shown in Fig. 10.V. CONCLUSION
In conclusion, we have analyzed the depth resolution for various ADSS frameworks under fixedconstrained resources. To evaluate the system performance of ADSS, we have considered system parameters including the number of cameras, the number of pixels, pixel size, and focal length. From the computational simulation, it is seen that the depth resolution in ADSS system can be improved with the large number of cameras and the large distance between optical axis and point source plane. The proposed analysis method may be one of a promising tool to design a practical ADSS system.

[FIG. 1.] Pickup process of ADSS system.

[FIG. 2.] Digital reconstruction process of ADSS system.

[FIG. 3.] Framework for ADSS with K stereo cameras.

[FIG. 4.] Calculation diagram of pixel position for each camera of the first point source.

[FIG. 5.] Calculation diagram of unresolvable pixel area.

[TABLE 1.] Experimental conditions for Monte Carlo simulation

[FIG. 6.] Simulation results according to the number of cameras with various distances between optical axis and point source plane.

[FIG. 7.] Simulation results according to the focal length of imaging lens. (a) When x=0 mm (b) When x=100 mm.

[FIG. 8.] Simulation results according to the pixel size of camera. (a) When x=0 mm (b) When x=100 mm.

[FIG. 9.] Simulation results according to the moving range of stereo camera. (a) When x=0 mm (b) When x=100 mm.

[FIG. 10.] Simulation results according to the total pixel number of stereo camera. (a) When x=0 mm (b) When x=100 mm.