OSA's Digital Library

Optics Express

Optics Express

  • Editor: C. Martijn de Sterke
  • Vol. 18, Iss. 25 — Dec. 6, 2010
  • pp: 25573–25583
« Show journal navigation

3D integral imaging display by smart pseudoscopic-to-orthoscopic conversion (SPOC)

H. Navarro, R. Martínez-Cuenca, G. Saavedra, M. Martínez-Corral, and B. Javidi  »View Author Affiliations


Optics Express, Vol. 18, Issue 25, pp. 25573-25583 (2010)
http://dx.doi.org/10.1364/OE.18.025573


View Full Text Article

Acrobat PDF (1297 KB)





Browse Journals / Lookup Meetings

Browse by Journal and Year


   


Lookup Conference Papers

Close Browse Journals / Lookup Meetings

Article Tools

Share
Citations

Abstract

Previously, we reported a digital technique for formation of real, non-distorted, orthoscopic integral images by direct pickup. However the technique was constrained to the case of symmetric image capture and display systems. Here, we report a more general algorithm which allows the pseudoscopic to orthoscopic transformation with full control over the display parameters so that one can generate a set of synthetic elemental images that suits the characteristics of the Integral-Imaging monitor and permits control over the depth and size of the reconstructed 3D scene.

© 2010 OSA

1. Introduction

Integral imaging (InI) is a three-dimensional (3D) imaging technique that works with incoherent light and provides with auto-stereoscopic images that can be observed without the help of special glasses. InI is the natural consequence of applying the modern technology to the old concept of Integral Photography (IP) which was proposed by Lipmann in 1908 [1

1. G. Lippmann, “Epreuves reversibles donnant la sensation du relief,” J. Phys. 7, 821–825 (1908).

]. Although InI concept was initially intended for the capture and display of 3D pictures or movies, in the last decade it has been proposed for other interesting applications. Examples of such applications are the digital reconstruction of spatially incoherent 3D scenes [2

2. S.-H. Hong, J.-S. Jang, and B. Javidi, “Three-dimensional volumetric object reconstruction using computational integral imaging,” Opt. Express 12(3), 483–491 (2004). [CrossRef] [PubMed]

,3

3. J.-H. Park, K. Hong, and B. Lee, “Recent progress in three-dimensional information processing based on integral imaging,” Appl. Opt. 48(34), H77–H94 (2009). [CrossRef] [PubMed]

], the visualization of partially occluded objects [4

4. B. Javidi, R. Ponce-Díaz, and S.-H. Hong, “Three-dimensional recognition of occluded objects by using computational integral imaging,” Opt. Lett. 31(8), 1106–1108 (2006). [CrossRef] [PubMed]

,5

5. B. Heigl, R. Koch, M. Pollefeys, J. Denzler, and L. Van Gool, “Plenoptic Modeling and Rendering from Image sequences taken by hand-held Camera,” Proc. DAGM, 94–101 (1999).

], or the sensing and recognition of dark scenes [6

6. S. Yeom, B. Javidi, and E. Watson, “Photon counting passive 3D image sensing for automatic target recognition,” Opt. Express 13(23), 9310–9330 (2005). [CrossRef] [PubMed]

].

One problem encountered with InI for 3D display applications is the pseudoscopic (or depth reversed) nature of the displayed images when the captured elemental images do not receive pixel pre-processing. Okano et al. were the first to propose a digital method to display orthoscopic scenes [7

7. F. Okano, H. Hoshino, J. Arai, and I. Yuyama, “Real-time pickup method for a three-dimensional image based on integral photography,” Appl. Opt. 36(7), 1598–1603 (1997). [CrossRef] [PubMed]

]. Although very simple and efficient, Okano’s algorithm has the weakness that it provides only virtual reconstructions, i.e. the 3D scene appears inside the monitor. Recently, we reported a method for the calculation of a set of synthetic elemental images (SEIs) that permit orthoscopic, real (or floating outside the monitor) reconstruction of the 3D scene [8

8. M. Martínez-Corral, B. Javidi, R. Martínez-Cuenca, and G. Saavedra, “Formation of real, orthoscopic integral images by smart pixel mapping,” Opt. Express 13(23), 9175–9180 (2005). [CrossRef] [PubMed]

]. Our algorithm, referred to as the Smart PIxel Mapping (SPIM), was however limited since it allows only a fixed position for the reconstructed scene. Also, the number of microlenses and their pitch cannot be changed. After the work reported in [8

8. M. Martínez-Corral, B. Javidi, R. Martínez-Cuenca, and G. Saavedra, “Formation of real, orthoscopic integral images by smart pixel mapping,” Opt. Express 13(23), 9175–9180 (2005). [CrossRef] [PubMed]

], research has been performed with the aim of overcoming the limitations of SPIM and/or taking profit of its capabilities such as the work reported by Shin et al. [9

9. D.-H. Shin, C.-W. Tan, B.-G. Lee, J.-J. Lee, and E.-S. Kim, “Resolution-enhanced three-dimensional image reconstruction by use of smart pixel mapping in computational integral imaging,” Appl. Opt. 47(35), 6656–6665 (2008). [CrossRef] [PubMed]

], were they propose a novel computational method based on SPIM for the improvement of the resolution of reconstructed scenes over long distances. Based on SPIM and using sub-image transformation process, a computational scheme was reported for removing occlusion in partially-occluded objects [10

10. M. Zhang, Y. Piao, and E.-S. Kim, “Occlusion-removed scheme using depth-reversed method in computational integral imaging,” Appl. Opt. 49(14), 2571–2580 (2010). [CrossRef]

,11

11. T.-Ch. Wei, D.-H. Shin, and B.-G. Lee, “Resolution-enhanced reconstruction of 3D object using depth-reversed elemental images for partially occluded object recognition,” J. Opt. Soc. Korea 13(1), 139–145 (2009). [CrossRef]

]. Also, there has been proposed modified versions of SPIM with the aim producing sets of synthetic elemental images created for displaying orthoscopic 3D images with depth control [12

12. D.-H. Shin, B.-G. Lee, and E.-S. Kim, “Modified smart pixel mapping method for displaying orthoscopic 3D images in integral imaging,” Opt. Lasers Eng. 47(11), 1189–1194 (2009). [CrossRef]

,13

13. J. Arai, H. Kawai, M. Kawakita, and F. Okano, “Depth-control method for integral imaging,” Opt. Lett. 33(3), 279–281 (2008). [CrossRef] [PubMed]

]. Other research have pursued the depth and scale control, but with no influence over the pseudoscopic nature of the reconstruction [14

14. D.-Ch. Hwang, J.-S. Park, S.-Ch. Kim, D.-H. Shin, and E.-S. Kim, “Magnification of 3D reconstructed images in integral imaging using an intermediate-view reconstruction technique,” Appl. Opt. 45(19), 4631–4637 (2006). [CrossRef] [PubMed]

].

In this paper we report an updated version of SPIM that permits the calculation of new sets of SEIs which are fully adapted to the display monitor characteristics. Specifically, this new pixel-mapping algorithm –denoted here as the Smart Pseudoscopic-to-Orthoscopic Conversion (SPOC), permits us to select the display parameters such as the pitch, focal length and size of the microlens array (MLA), the depth position and size of the reconstructed images, and even the geometry of the MLA.

To present the new approach, we have organized this paper as following. In Section 2, we explain the principles of SPOC and develop the corresponding mathematical formulation. In Section 3, we revisit two classical algorithms for the pseudoscopic-to-orthoscopic (PO) conversion and demonstrate that they can be considered as particular cases of the SPOC algorithm. Section 4 is devoted to the computer and experimental verifications of the SPOC. Finally in Section 5, we summarize the main achievements of this reported research. Acronyms in this paper are listed in Table 1.

2. The smart pseudoscopic-to-orthoscopic conversion algorithm

Now in this paper we have taken profit from the additional possibilities of SPIM and have developed a more flexible digital method that allows the calculation of a new set of SEIs to be used in a display configuration that can be essentially different from the one used in the capture. We have named the new algorithm “SPOC”. It allows the calculation of SEIs ready to be displayed in an InI monitor in which the pitch, the microlenses focal length, the number of pixels per elemental cell, the depth position of the reference plane, and even the grid geometry of the MLA can be selected to fit the conditions of the display architecture.

Table 1. List of acronyms

table-icon
View This Table

The SPOC algorithm is relatively simple. One has to calculate the SEIs that are obtained in the simulated experiment schematized in Fig. 1
Fig. 1 Calculation of the synthetic integral image. The pixel of the synthetic integral image (dotted blue line) stores the same value as the pixel of the captured integral image.
. The algorithm can be explained as the result of the application in cascade of three processes: the simulated display, the synthetic capture and the homogeneous scaling.

In the simulated display we use as the input for the algorithm, the set of elemental images captured experimentally. The pitch, the gap and the focal length of the MLA are equal to those used in the experimental capture. This simulated display permits the reconstruction of the original 3D scene in the same position and with the same scale.

The second process is the synthetic capture, which is done through an array of pinholes (PA). To give our algorithm the maximum generality, for the synthetic capture: (i) we place the synthetic PA at an arbitrary distance D from the display MLA; (ii) we assign arbitrary pitch, p S, and gap, g S, to the synthetic PA –note that this selection will determine also the size of the final image; and (iii) we fix, also arbitrarily, the number of pixels, n S, per synthetic elemental image and the total number of elemental images, N S.

Note that the value of parameter d S determines the position of the reference plane of the image displayed by the InI display monitor. A smart selection of d S will permit, when displaying the SEIs in an actual InI display monitor, the observation of either orthoscopic real or virtual 3D images. A positive value of d S corresponds to a floating real 3D image. A negative value corresponds to a virtual reconstruction.

The third step, the homogeneous scaling, is intended to adapt the size (scale) of the SEIs to the final InI monitor. In this step, both the pitch and the gap are scaled by the same factor. The only constraint in this step is that the final SEIs must be ready to be used in a realistic InI monitor and therefore the value of the scaled gap should be equal to the focal length of the monitor microlenses.

For simplicity, we have drawn in Fig. 1 the one-dimensional case. The extrapolation of the forthcoming theory to two dimensions is trivial. There exist, however, a case of great interest in which the extrapolation is not trivial. We refer to the case in which the geometry of synthetic array is different from that of the image capture stage. This happens, for example, when one of the arrays is rectangular and the other hexagonal.

IjmS=IilD
(7)

3. Revisiting two typical cases of the pseudoscopic to orthoscopic (PO) conversion

To demonstrate the generality of this proposal, we revisit two classical algorithms for the PO conversion and demonstrate that they can be considered as particular cases of the SPOC algorithm.

3.1. The method proposed by Okano et al

A smart and simple method for the PO conversion was reported by Okano and associates [7

7. F. Okano, H. Hoshino, J. Arai, and I. Yuyama, “Real-time pickup method for a three-dimensional image based on integral photography,” Appl. Opt. 36(7), 1598–1603 (1997). [CrossRef] [PubMed]

], who proposed to rotate each elemental image by 180° around the center of the elemental cell. Then, the rotated elemental images are displayed at a distance g S = g D -2f 2 /(d D -f) from a MLA similar to the one used in the capture. This procedure permits the reconstruction of virtual, orthoscopic 3D scenes. Note however that gSgD, and therefore the final reconstructed image is slightly distorted since it has shrunk in the axial direction.

To reproduce Okano’s method one simply has to use as the input of the SPOC algorithm the following values (see Fig. 2
Fig. 2 Scheme for the calculation of the Okano’s synthetic integral image.
): D=0 (and, thereforedS=dD), gS=gD, pS=pD and nS=nD.

Introducing such values into Eqs. (4) and (6), one finds
ijm=j
(8)
and
ljm=Round[gDdSnDim+nDgSdSj]=m
(9)
Note that SPOC result is, however, slightly different from the one reported by Okano and his colleagues. While in their case fS=fD but gSgD, in our case gS=gD and fS=gS. This fact gives the SPOC a slight advantage, since it permits the reconstruction of 3D scenes without any distortion, i.e. with the same magnification in the axial and in the lateral direction, and also free of facet braiding [15

15. H. Navarro, R. Martínez-Cuenca, A. Molina-Martín, M. Martínez-Corral, G. Saavedra, and B. Javidi, “Method to remedy image degradations due to facet braiding in 3D integral imaging monitors,” J. Display Technol. 6(10), 404–411 (2010). [CrossRef]

]. On the other hand, one can use SPOC to produce, in addition to the PO conversion, other changes to adapt the integral image to the display grid geometry.

3.2. The symmetric case

This is the case for which the SPIM algorithm was created. Again, g S = g D, p S = p D and n S = n D, but now D = 2d D (and therefore and d S = d D), besides
nD=nS=2dDgD
(10)
This leads to the result
ijm=jm
(11)
ljm=m
(12)
Note that these equations are the same as the ones reported in [8

8. M. Martínez-Corral, B. Javidi, R. Martínez-Cuenca, and G. Saavedra, “Formation of real, orthoscopic integral images by smart pixel mapping,” Opt. Express 13(23), 9175–9180 (2005). [CrossRef] [PubMed]

], although with different appearance due to the choice of different label criteria.

4. Demonstration of the SPOC and experimental results

To demonstrate the versatility of the SPOC, we apply it to generate SEIs ready to be displayed in display architectures which are very different from the parameters used in capture stage. For the capture of the elemental images, we prepared over a black background a 3D scene composed by a doll (a cook preparing paella, see Fig. 3
Fig. 3 Scheme of the experimental set up for the acquisition of the set of elemental images of a 3D scene.
).

For the acquisition of the elemental images, instead of using an array of digital cameras, we used the so-called synthetic aperture method [5

5. B. Heigl, R. Koch, M. Pollefeys, J. Denzler, and L. Van Gool, “Plenoptic Modeling and Rendering from Image sequences taken by hand-held Camera,” Proc. DAGM, 94–101 (1999).

,16

16. J. S. Jang and B. Javidi, “Three-dimensional synthetic aperture integral imaging,” Opt. Lett. 27(13), 1144–1146 (2002). [CrossRef]

], in which all the elemental images are picked up with only one digital camera that is mechanically translated. The digital camera was focused at a distance of one meter. The camera parameters were fixed to f=10mm and f/#=22. The depth of field was large enough to allow obtaining sharp pictures of the doll, which was placed at a distance d=203mm from the camera. The gap for the capture was, then, g=10mm. We obtained a set of NH=17×NV=11 images with pitch PH=22mm andPV=14mm. Note that the pitch is slightly smaller than the size of the CMOS sensor (22.2 mm x 14.8 mm), thus we cropped slightly any captured picture in order to remove the outer parts. By this way, we could compose the integral image consisting of 17H ×11V elemental images of 22×14 mm and nH=2256×nV=1504  pixels each.

In Fig. 4
Fig. 4 Subset of the elemental images obtained experimentally. These elemental images are the input for the SPOC algorithm.
, we show a portion of the set of elemental images obtained after the capture experiment and the cropping of the pictures.

For the calculation of the set of SEIs, and with the aim of reducing the computing time, we first resized the elemental images to nDH=251px , and nDV=161px .. Then we fixed the synthetic parameters to: dS=dD/2=101.5mm, gS=3.75mm, pSH=pSV=1.25mm, nSH=nSV=19px and NSH=NSV=151 microlenses. In Fig. 5
Fig. 5 (a)Collection of 151×151 SEIs obtained after the application of the SPOC algorithm; (b) enlarged view of central SEIs.
we show the set of calculated SEIs. Note that we have increased substantially the number of elemental images, which now are square and arranged in square grid.

Finally we applied the third step of SPOC, and scaled the corresponding parameter by factor of 1.25, so that gS=3.0mm, pSH=pSV=1.0mmand therefore dS=81.2mm. With this final set of SEIs we performed two visualization experiments. One simulated with the computer, the second real in our laboratory.

For the simulated visualization experiment, the calculations were done assuming a virtual observer placed at a distance L=700mm from the MLA. The visualization calculations were performed following the algorithm described in [15

15. H. Navarro, R. Martínez-Cuenca, A. Molina-Martín, M. Martínez-Corral, G. Saavedra, and B. Javidi, “Method to remedy image degradations due to facet braiding in 3D integral imaging monitors,” J. Display Technol. 6(10), 404–411 (2010). [CrossRef]

]. In Fig. 6
Fig. 6 Two Perspectives of the 3D reconstructed scene, as seen by an observer placed at a distanceL=700mm. In the video (Media 1) we show the movie built with the frames obtained with the visualization algorithm.
we show the result of the visualization simulation.

Next, we performed the optical visualization experiment. To this end, first we printed the SEIs in photographic paper with a high-resolution inkjet printer. Our InI display monitor was equipped with by an array of 151×151 microlenses, arranged in square grid, with pitch p=1.0mm and f=3.0mm. Then we placed the SEIs at a distance g=3.0mm from the MLA (see Fig. 7
Fig. 7 Experimental setup for the observation of the InI monitor. After displacing horizontally the camera in steps of 10 mm we recorded 20 different perspectives of the displayed 3D scene.
).

Up to 20 different perspectives of the displayed 3D image were captured with a digital camera placed at L=700mm. The camera was displaced in the horizontal direction in steps of 10mm. The captured perspectives are shown in Fig. 8
Fig. 8 Two perspectives of the 3D scene displayed in the real experiment. In the video (Media 2) we show the movie built with the perspectives photographed with the digital camera.
.

We have demonstrated that the SPOC permits to create from a low number of elemental images, a new set of SEIs ready to be displayed in an InI monitor equipped with a MLA composed by a much higher number of microlenses. The displayed image is orthoscopic, and is displayed at a shorter distance from the monitor.

Next, and as the proof of the flexibility of the SPOC algorithm, we calculated the SEIs for a display geometry that is essentially different from the one used in the capture stage, but that is very common in the 3D display applications [17

17. J. Arai, H. Kawai, and F. Okano, “Microlens arrays for integral imaging system,” Appl. Opt. 45(36), 9066–9078 (2006). [CrossRef] [PubMed]

]. We refer to the case in which the display microlenses are arranged in a hexagonal grid. For the application of the SPOC algorithm, we considered microlenses with diameter ϕS=1.0mm and focal length fS=3.0mm. Besides, we fixed the depth distance to the MLA as dS=20.3mm. We applied the SPOC to calculate up to 24462 hexagonal synthetic elemental images (the dimensions of the hexagonal MLA were 151×152mm), which are shown in Fig. 9
Fig. 9 (a)Collection of hexagonal elemental images obtained after the application of the SPOC algorithm; (b) enlarged view of some central SEIs.
.

For the simulated visualization experiment the calculations were done assuming a virtual observer placed at a distance L=700mm from the MLA. In Fig. 10
Fig. 10 Two Perspectives of the reconstructed scene, as seen by an observer placed at a distanceL=700mm. In the video (Media 3) we show the movie built with the frames obtained with the visualization algorithm.
we show the result of the visualization simulation. The degradations that are observed in the reconstructed image are due to the fact that there is no possibility of perfect matching between the microlenses arranged in hexagonal grid and the pixels of the matrix display, which are arranged in rectangular grid.

Also in this case we performed optical visualization experiments. We printed the SEIs on photographic paper. The InI monitor was equipped with an array of microlenses, of diameter ϕ=1.0mm and f=3.0mm arranged in hexagonal grid. Then we placed the SEIs at a distance g=3.0mm from the MLA (see Fig. 11
Fig. 11 Experimental setup for the observation of the hexagonal InI monitor. After displacing horizontally the camera in steps of 10 mm, we recorded 35 different perspectives of the displayed 3D scene.
).

Up to 35 different perspectives of the displayed 3D image were captured with a digital camera placed at L=700mm. The camera was moved in the horizontal direction in steps of

10 nm. The captured perspectives are shown in Fig. 12
Fig. 12 Two perspectives of the 3D scene displayed in the real hexagonal experiment. In the video (Media 4) we show the movie built with the perspectives photographed with the digital camera.
.

One interesting issue to discuss here is how the SPOC affects the resolution of reconstructed images. There is no one single response, since it depends of the algorithm parameter. Since it is not possible to increment the image bandwidth through rearrangement of pixel information, no increment of image resolution is possible with SPOC. If we do not want to lose resolution, it is necessary select carefully the parameters of the algorithm.

A different issue is the resolution of the reconstructed image as observed when the observer eye looks at the display. As explained in [15

15. H. Navarro, R. Martínez-Cuenca, A. Molina-Martín, M. Martínez-Corral, G. Saavedra, and B. Javidi, “Method to remedy image degradations due to facet braiding in 3D integral imaging monitors,” J. Display Technol. 6(10), 404–411 (2010). [CrossRef]

], such viewing resolution is mainly determined by the microlenses pitch. In this case, we can state that the proper use of SPOC can help to increment significantly the viewing resolution.

5. Conclusions

We have demonstrated the SPOC algorithm which allows full control over the optical display parameters in InI monitors. Specifically, we have shown that from a given collection of elemental images, one can create a new set of SEIs ready to be displayed in an InI monitor in which the pitch, the microlenses focal length, the number of pixels per elemental cell, the depth position of the reference plane, and even the grid geometry of the MLA can be selected to fit the conditions of the display architecture.

Acknowledgements

References and Links

1.

G. Lippmann, “Epreuves reversibles donnant la sensation du relief,” J. Phys. 7, 821–825 (1908).

2.

S.-H. Hong, J.-S. Jang, and B. Javidi, “Three-dimensional volumetric object reconstruction using computational integral imaging,” Opt. Express 12(3), 483–491 (2004). [CrossRef] [PubMed]

3.

J.-H. Park, K. Hong, and B. Lee, “Recent progress in three-dimensional information processing based on integral imaging,” Appl. Opt. 48(34), H77–H94 (2009). [CrossRef] [PubMed]

4.

B. Javidi, R. Ponce-Díaz, and S.-H. Hong, “Three-dimensional recognition of occluded objects by using computational integral imaging,” Opt. Lett. 31(8), 1106–1108 (2006). [CrossRef] [PubMed]

5.

B. Heigl, R. Koch, M. Pollefeys, J. Denzler, and L. Van Gool, “Plenoptic Modeling and Rendering from Image sequences taken by hand-held Camera,” Proc. DAGM, 94–101 (1999).

6.

S. Yeom, B. Javidi, and E. Watson, “Photon counting passive 3D image sensing for automatic target recognition,” Opt. Express 13(23), 9310–9330 (2005). [CrossRef] [PubMed]

7.

F. Okano, H. Hoshino, J. Arai, and I. Yuyama, “Real-time pickup method for a three-dimensional image based on integral photography,” Appl. Opt. 36(7), 1598–1603 (1997). [CrossRef] [PubMed]

8.

M. Martínez-Corral, B. Javidi, R. Martínez-Cuenca, and G. Saavedra, “Formation of real, orthoscopic integral images by smart pixel mapping,” Opt. Express 13(23), 9175–9180 (2005). [CrossRef] [PubMed]

9.

D.-H. Shin, C.-W. Tan, B.-G. Lee, J.-J. Lee, and E.-S. Kim, “Resolution-enhanced three-dimensional image reconstruction by use of smart pixel mapping in computational integral imaging,” Appl. Opt. 47(35), 6656–6665 (2008). [CrossRef] [PubMed]

10.

M. Zhang, Y. Piao, and E.-S. Kim, “Occlusion-removed scheme using depth-reversed method in computational integral imaging,” Appl. Opt. 49(14), 2571–2580 (2010). [CrossRef]

11.

T.-Ch. Wei, D.-H. Shin, and B.-G. Lee, “Resolution-enhanced reconstruction of 3D object using depth-reversed elemental images for partially occluded object recognition,” J. Opt. Soc. Korea 13(1), 139–145 (2009). [CrossRef]

12.

D.-H. Shin, B.-G. Lee, and E.-S. Kim, “Modified smart pixel mapping method for displaying orthoscopic 3D images in integral imaging,” Opt. Lasers Eng. 47(11), 1189–1194 (2009). [CrossRef]

13.

J. Arai, H. Kawai, M. Kawakita, and F. Okano, “Depth-control method for integral imaging,” Opt. Lett. 33(3), 279–281 (2008). [CrossRef] [PubMed]

14.

D.-Ch. Hwang, J.-S. Park, S.-Ch. Kim, D.-H. Shin, and E.-S. Kim, “Magnification of 3D reconstructed images in integral imaging using an intermediate-view reconstruction technique,” Appl. Opt. 45(19), 4631–4637 (2006). [CrossRef] [PubMed]

15.

H. Navarro, R. Martínez-Cuenca, A. Molina-Martín, M. Martínez-Corral, G. Saavedra, and B. Javidi, “Method to remedy image degradations due to facet braiding in 3D integral imaging monitors,” J. Display Technol. 6(10), 404–411 (2010). [CrossRef]

16.

J. S. Jang and B. Javidi, “Three-dimensional synthetic aperture integral imaging,” Opt. Lett. 27(13), 1144–1146 (2002). [CrossRef]

17.

J. Arai, H. Kawai, and F. Okano, “Microlens arrays for integral imaging system,” Appl. Opt. 45(36), 9066–9078 (2006). [CrossRef] [PubMed]

OCIS Codes
(100.6890) Image processing : Three-dimensional image processing
(110.4190) Imaging systems : Multiple imaging
(110.6880) Imaging systems : Three-dimensional image acquisition
(120.2040) Instrumentation, measurement, and metrology : Displays

ToC Category:
Imaging Systems

History
Original Manuscript: September 22, 2010
Revised Manuscript: October 27, 2010
Manuscript Accepted: October 27, 2010
Published: November 22, 2010

Citation
H. Navarro, R. Martínez-Cuenca, G. Saavedra, M. Martínez-Corral, and B. Javidi, "3D integral imaging display by smart 
pseudoscopic-to-orthoscopic conversion (SPOC)," Opt. Express 18, 25573-25583 (2010)
http://www.opticsinfobase.org/oe/abstract.cfm?URI=oe-18-25-25573


Sort:  Author  |  Year  |  Journal  |  Reset  

References

  1. G. Lippmann, “Epreuves reversibles donnant la sensation du relief,” J. Phys. 7, 821–825 (1908).
  2. S.-H. Hong, J.-S. Jang, and B. Javidi, “Three-dimensional volumetric object reconstruction using computational integral imaging,” Opt. Express 12(3), 483–491 (2004). [CrossRef] [PubMed]
  3. J.-H. Park, K. Hong, and B. Lee, “Recent progress in three-dimensional information processing based on integral imaging,” Appl. Opt. 48(34), H77–H94 (2009). [CrossRef] [PubMed]
  4. B. Javidi, R. Ponce-Díaz, and S.-H. Hong, “Three-dimensional recognition of occluded objects by using computational integral imaging,” Opt. Lett. 31(8), 1106–1108 (2006). [CrossRef] [PubMed]
  5. B. Heigl, R. Koch, M. Pollefeys, J. Denzler, and L. Van Gool, “Plenoptic Modeling and Rendering from Image sequences taken by hand-held Camera,” Proc. DAGM, 94–101 (1999).
  6. S. Yeom, B. Javidi, and E. Watson, “Photon counting passive 3D image sensing for automatic target recognition,” Opt. Express 13(23), 9310–9330 (2005). [CrossRef] [PubMed]
  7. F. Okano, H. Hoshino, J. Arai, and I. Yuyama, “Real-time pickup method for a three-dimensional image based on integral photography,” Appl. Opt. 36(7), 1598–1603 (1997). [CrossRef] [PubMed]
  8. M. Martínez-Corral, B. Javidi, R. Martínez-Cuenca, and G. Saavedra, “Formation of real, orthoscopic integral images by smart pixel mapping,” Opt. Express 13(23), 9175–9180 (2005). [CrossRef] [PubMed]
  9. D.-H. Shin, C.-W. Tan, B.-G. Lee, J.-J. Lee, and E.-S. Kim, “Resolution-enhanced three-dimensional image reconstruction by use of smart pixel mapping in computational integral imaging,” Appl. Opt. 47(35), 6656–6665 (2008). [CrossRef] [PubMed]
  10. M. Zhang, Y. Piao, and E.-S. Kim, “Occlusion-removed scheme using depth-reversed method in computational integral imaging,” Appl. Opt. 49(14), 2571–2580 (2010). [CrossRef]
  11. T.-Ch. Wei, D.-H. Shin, and B.-G. Lee, “Resolution-enhanced reconstruction of 3D object using depth-reversed elemental images for partially occluded object recognition,” J. Opt. Soc. Korea 13(1), 139–145 (2009). [CrossRef]
  12. D.-H. Shin, B.-G. Lee, and E.-S. Kim, “Modified smart pixel mapping method for displaying orthoscopic 3D images in integral imaging,” Opt. Lasers Eng. 47(11), 1189–1194 (2009). [CrossRef]
  13. J. Arai, H. Kawai, M. Kawakita, and F. Okano, “Depth-control method for integral imaging,” Opt. Lett. 33(3), 279–281 (2008). [CrossRef] [PubMed]
  14. D.-Ch. Hwang, J.-S. Park, S.-Ch. Kim, D.-H. Shin, and E.-S. Kim, “Magnification of 3D reconstructed images in integral imaging using an intermediate-view reconstruction technique,” Appl. Opt. 45(19), 4631–4637 (2006). [CrossRef] [PubMed]
  15. H. Navarro, R. Martínez-Cuenca, A. Molina-Martín, M. Martínez-Corral, G. Saavedra, and B. Javidi, “Method to remedy image degradations due to facet braiding in 3D integral imaging monitors,” J. Display Technol. 6(10), 404–411 (2010). [CrossRef]
  16. J. S. Jang and B. Javidi, “Three-dimensional synthetic aperture integral imaging,” Opt. Lett. 27(13), 1144–1146 (2002). [CrossRef]
  17. J. Arai, H. Kawai, and F. Okano, “Microlens arrays for integral imaging system,” Appl. Opt. 45(36), 9066–9078 (2006). [CrossRef] [PubMed]

Cited By

Alert me when this paper is cited

OSA is able to provide readers links to articles that cite this paper by participating in CrossRef's Cited-By Linking service. CrossRef includes content from more than 3000 publishers and societies. In addition to listing OSA journal articles that cite this paper, citing articles from other participating publishers will also be listed.

Supplementary Material


» Media 1: AVI (581 KB)     
» Media 2: AVI (389 KB)     
» Media 3: AVI (283 KB)     
» Media 4: AVI (592 KB)     

« Previous Article  |  Next Article »

OSA is a member of CrossRef.

CrossCheck Deposited