Compact single-shot metalens depth sensors inspired by eyes of jumping spiders
Contributed by Federico Capasso, September 24, 2019 (sent for review July 16, 2019; reviewed by David Stork and Jingyi Yu)
Significance
Nature provides diverse solutions to passive visual depth sensing. Evolution has produced vision systems that are highly specialized and efficient, delivering depth-perception capabilities that often surpass those of existing artificial depth sensors. Here, we learn from the eyes of jumping spiders and demonstrate a metalens depth sensor that shares the compactness and high computational efficiency of its biological counterpart. Our device combines multifunctional metalenses, ultrathin nanophotonic components that control light at a subwavelength scale, and efficient computations to measure depth from image defocus. Compared with previous passive artificial depth sensors, our bioinspired design is lightweight, single-shot, and requires a small amount of computation. The integration of nanophotonics and efficient computation establishes a paradigm for design in computational sensing.
Abstract
Jumping spiders (Salticidae) rely on accurate depth perception for predation and navigation. They accomplish depth perception, despite their tiny brains, by using specialized optics. Each principal eye includes a multitiered retina that simultaneously receives multiple images with different amounts of defocus, and from these images, distance is decoded with relatively little computation. We introduce a compact depth sensor that is inspired by the jumping spider. It combines metalens optics, which modifies the phase of incident light at a subwavelength scale, with efficient computations to measure depth from image defocus. Instead of using a multitiered retina to transduce multiple simultaneous images, the sensor uses a metalens to split the light that passes through an aperture and concurrently form 2 differently defocused images at distinct regions of a single planar photosensor. We demonstrate a system that deploys a 3-mm-diameter metalens to measure depth over a 10-cm distance range, using fewer than 700 floating point operations per output pixel. Compared with previous passive depth sensors, our metalens depth sensor is compact, single-shot, and requires a small amount of computation. This integration of nanophotonics and efficient computation brings artificial depth sensing closer to being feasible on millimeter-scale, microwatts platforms such as microrobots and microsensor networks.
Sign up for PNAS alerts.
Get alerts for new articles, or get an alert when an article is cited.
Visual depth sensors combine cameras, computational algorithms, and sometimes light sources to sense the 3-dimensional shapes of surrounding objects and scenes. Lidar systems (1), time-of-flight cameras (2–8), and structured lighting systems (9, 10) are examples of depth sensors that use active light sources, whereas binocular stereo systems (11) and light-field cameras (12–14) are examples that are passive, relying solely on the ambient light that happens to be available. These approaches have found widespread use on autonomous vehicles, drones, mobile phones, and many other platforms. However, they require either active lighting or iterative computation and optimization, and are thus not well suited to low-power platforms, such as mobile sensor networks and robotic insects (15–17), which impose much more severe constraints on size, weight, and power consumption.
Alternative methods that utilize optical defocus to measure depth have been demonstrated to potentially greatly reduce the amount of depth computation and require no active lighting (18, 19). These algorithms (18–23) compute depth by comparing 2 differently defocused images of the same scene and produce a depth map, comprising a depth value at each pixel. However, one major challenge with this method is the optics. With conventional optical components, capturing 2 differently defocused images usually requires making physical changes to the optical system, such as reducing or enlarging its aperture (20, 21, 23) or deforming its lens (19). This not only adds significant complexity to the system control, but also fundamentally limits the depth-sensing performance by introducing unwanted delays and motion artifacts. Some previous algorithms use look-up tables (20) or iterative methods (22) in their framework to measure depth. However, these methods are hard to implement in a differentiable manner and rely on exhaustive search, instead of using gradient-based search methods, to determine the required parameters.
To address these challenges, we introduce the metalens depth sensor. It is compact, static, single-shot, and requires low computational power. Thanks to the versatile wavefront-shaping capability of metalenses, ultrathin nanophotonic components that can tailor arbitrary optical wavefront at a subwavelength scale, our device can simultaneously capture 2 differently defocused images through the same aperture without having to make physical changes to the optical system. It avoids the artifacts usually incurred by reimaging over time while changing a camera’s optics and can potentially improve the depth sensor’s time resolution. Besides, the image-processing algorithm is completely differentiable, which enables data-driven, gradient-based calibration of the computational parameters compared to the nondifferentiable methods (20, 22).
The working principle is inspired by the eyes of jumping spiders (Salticidae), which use defocus to succeed at sensing depth, despite the fact that their brains are about as small as poppy seeds (24). Each of the spider’s principal eyes includes a specialized structure (25) with stacked translucent retinae that simultaneously observe the world with different amounts of optical defocus (Fig. 1A). Behavioral experiments have shown that the input from 1 principal eye suffices for a jumping spider to sense depth accurately enough to leap onto prey from distances of several body lengths and that depth perception can be predictably manipulated by changing the ambient light spectrum in a way that distorts the optical defocus (26).
Fig. 1.

Inspired by the specialized compact optical structure of the jumping spider’s principal eye, we propose to use metasurface technology (27–34) to simultaneously collect a pair of differently defocused images using the 2 halves of a single planar photosensor (Fig. 1B). Metasurfaces are ultrathin planar optical components consisting of subwavelength-spaced nanostructures patterned at an interface (27). By engineering the shape of individual nanostructures, one can control the phase, amplitude, and polarization of the transmitted wavefront at subwavelength scales, allowing multiple functions to be multiplexed within a single device. Metasurfaces have enabled a variety of optical devices with capabilities that surpass those of conventional refractive or diffractive elements, ranging from high-performance imaging lenses (metalenses) (29, 35) to novel polarization holograms (36). In our prototype sensor, we encode 2 complementary lens phase profiles with distinct focal lengths and lateral offsets on a shared aperture in a single metalens by spatial multiplexing (28). In this way, 2 differently defocused images can be captured simultaneously side by side on the photosensor in a single shot. We design the metalens focal lengths together using a depth-reconstruction algorithm, so that accurate depth maps can be computed from the 2 simultaneous images with calculations that are spatially localized and few in number—i.e., depth computations for each image pixel involve only a small spatial neighborhood of pixels and require no additional correspondence search after initial calibration.
Our prototype produces depth values over a range of 10 cm from single-shot measurements using a millimeter-scale metalens. Calculating depth at each output pixel of the depth map requires fewer than 700 floating point operations (FLOPs) and involves the digitized intensity values in only a 25 25 spatial neighborhood of pixels. This integration of nanophotonics and efficient computation brings artificial depth sensing closer to being feasible on millimeter-scale, microwatts platforms such as microrobots and microsensor networks.
Principle
We model the image formed on a photosensor as the convolution of the camera point spread function (PSF) with the magnified, all-in-focus object pattern as it would be observed with a pinhole camera. The camera PSF is the image captured on the photosensor when the object is a point light source. The width of the PSF depends on the optics and the distance Z between the object and the lens. For an ideal thin-lens camera, depicted in Fig. 2A, the PSF width is related to object distance by the thin-lens equation:where is the distance between the lens and the photosensor, is the radius of the entrance pupil, and is the in-focus distance—i.e., the distance for which the PSF width is equal to zero. (SI Appendix, section 1.1) On the right side of Eq. 1, all quantities but the object distance Z are known quantities determined by the optical system. Thus, for a calibrated camera, determining the PSF width is equivalent to measuring object distance .
[1]
Fig. 2.

The PSF width determines the amount of image blur. An object appears sharp when its distance is equal to the in-focus distance because then the PSF width is zero (under ray optics approximation). Conversely, when the object distance deviates from , the PSF width, , is nonzero, and the image is blurry. Recovering the PSF width (and thus depth) from a single blurry image is ill-posed without prior information about the underlying object pattern. However, when a second image of the same scene is captured with a different amount of blur, the width can be determined directly from the contrast change between the 2 images. One way to understand this is to assume that the PSFs can be approximated as Gaussian functions that depend on the PSF width :where is the pixel position on the photosensor. Gaussian functions have the property that partial derivatives with respect to width and location satisfyand because the defocused image is the convolution of the PSF and the all-in-focus object pattern (which does not depend on ), the same relationship between derivatives applies to the captured image:Eq. 4 indicates that (and thus depth Z through Eq. 1) can be determined directly from the spatial Laplacian of the image and the differential change of intensity with respect to varying PSF width (18, 19). The latter can be estimated via a finite difference, i.e., , where is the change of image intensity induced by a small, known variation of the PSF width (). According to Eq. 1, since in general no control can be made over the object distance , the only way to change the PSF width when shooting an object is to vary the parameters of the optical system, such as the sensor distance or the in-focus distance .
[2]
[3]
[4]
A jumping spider’s principal eye can use its transparent layered retinae to simultaneously measure the (minimally) 2 images that are required to compute the finite difference (Fig. 1A) because these retinae effectively capture images with different sensor distances . In contrast, we design a metalens that generates 2 images (, ) side by side (Fig. 1B), with the images being equivalent to images that are captured with different in-focus distances (, ) through the same pupil. We design the in-focus distances so that the difference in blur between the images, , is small and approximately differential. From these 2 images (Fig. 2 C, Left), we compute the per-pixel difference and the image Laplacian . The latter is obtained by convolving the averaged image with a Laplacian filter, denoted . To reduce the effects of sensor noise and optical nonidealities like vignetting, and are spatially convolved with a purposefully designed linear filter . To the lowest order, the filter is similar to a Gaussian filter that averages over neighboring pixels (SI Appendix, sections 1.3 and 2.2). The filtered results and are shown in Fig. 2 C, Center. Finally, we combine Eqs. 1 and 4 to calculate the depth at each pixel :with and being constants that are determined by the optics. The correctness of Eq. 5 follows from the fact that Eq. 4 still holds when values and are replaced by their filtered versions and .
[5]
In practice, even with filtering, random noise in the captured images (, ) results in errors in the measured depth . The error can be quantified in terms of the SD of the measured depth at each pixel, which can be approximated by the measurable quantitywith constants that are determined by the optics (SI Appendix, section 1.2). This measurable quantity can serve as an indicator of the reliability of the measured depth at each pixel . For convenience, we normalize the values of to the range and define this normalized value as the confidence . A higher confidence value at pixel location indicates a smaller value of and a more accurate depth measurement (SI Appendix, section 1.2). Physically, the confidence characterizes the expected accuracy of the measurement at each pixel : A larger confidence value at a pixel indicates a statistically smaller error in the depth measurement.
[6]
Since Eq. 5 calculates depth using simple, local calculations, it fails in regions of the images that have uniform intensity and thus no measurable contrast for and . To automatically identify the locations of these failures, we use the confidence score as a criterion, and we report depth only at pixels whose confidence is above a certain threshold. The choice of confidence threshold affects the depth resolution, which we define as the smallest depth difference that can be resolved within a certain confidence range. In this paper, with a confidence threshold of 0.5, we achieve a depth resolution of about of the object distance over the distance range (see Fig. 4B).
The complete sequence of calculations for the depth map and confidence map is depicted in Fig. 2C. For visualization, the depth map is thresholded by confidence to show only the depth at pixels where the latter is greater than 0.5.
Metalens Design and Characterization
The metalens is designed to incorporate phase profiles of 2 off-axis lenses with different in-focus distances on a shared aperture. For each off-axis lens, the required phase profile is an offset convex shape determined by in-focus distances , sensor distance , and transverse displacement of the image center (Fig. 2B):Here, indicates location on the metalens. The overall phase profile is achieved by spatially interleaving the 2, and , on the metalens at a subwavelength scale. The design specifications are in SI Appendix, section 2.1.
[7]
The required phase profile can be wrapped to (i.e., modulo ) without changing its functionality. Therefore, the key requirement for precise wavefront shaping is to control locally the phase between 0 and . Here, we use titanium dioxide () nanopillars as the building blocks for local phase control. Physically, the nanopillars function as truncated waveguides and impart a phase shift to the transmitted light. By varying the pillar width, one can tune their effective refractive index, and thus the phase shift. Fig. 3A shows that by changing the pillar width () from 90 to 190 nm, one can achieve 0 to phase coverage, while maintaining a high transmission efficiency. The nanopillars have a uniform height () of 600 nm and can be fabricated with a single-step lithography. The center-to-center distance () between the neighboring nanopillars is 230 nm, smaller than half the operating wavelength. This allows us to spatially interleave different phase profiles at a subwavelength scale, which is essential to eliminating unwanted higher-order diffractions.
Fig. 3.

The metalens is fabricated with a technique demonstrated by Devlin et al. (34). Fig. 3 B–D show the scanning electron microscope (SEM) images of a fabricated sample. The location of the imaged area on the metalens is in SI Appendix, Fig. S6. The phase wrapping introduces a discontinuity at locations where the phase profile equals an integer number of —i.e., the “zone” boundaries. This corresponds to an abrupt change of nanopillar arrangement, as shown in Fig. 3 B–D. The phase profiles of 2 off-axis lenses have different zone spacing and orientation, corresponding to the 2 nearly vertical boundaries and the diagonal boundary, respectively (Fig. 3B and SI Appendix, Fig. S6). The spatial multiplexing scheme is illustrated explicitly in Fig. 3C, with nanopillars belonging to different focusing profiles highlighted in different colors.
Results
We built a prototype metalens depth sensor by coupling the metalens with off-the-shelf components. The sensor’s current size, including mechanical components such as optical mounts, is 4 4 10 cm, but since the metalens is only 3 mm in diameter, the overall size of the assembled sensor could be reduced substantially with a purpose-built photosensor and housing. We paired a 10-nm bandpass filter with the metalens, which is designed for monochromatic operation at 532 nm. A rectangular aperture was placed in front of the metalens to limit the field of view and prevent the 2 images from overlapping. The blur change between the 2 images can be seen in Fig. 4A, which shows PSFs for each of the 2 images [, ] that were measured by using a green light-emitting diode (LED) mated to a 10-m-diameter pinhole and placed at different depths Z along the optical axis. The PSFs are more disc-like than Gaussian, and they are asymmetric due to off-axis chromatic aberration. (SI Appendix, Fig. S8 shows that this asymmetry disappears under monochromatic laser illumination.)
Fig. 4.

To suppress the effects of noise and imaging artifacts in images (, ), and to increase the number of high-confidence pixels in the output maps, we computed 9 separate depth and confidence maps using 9 instances of Eqs. 5 and 6 that have distinct and complementary spatial filters , and then we fused these 9 “channels” into 1. We also designed a calibration procedure that tuned the parameters simultaneously, using back-propagation and gradient descent (SI Appendix, section 3). In addition to being user-friendly, this end-to-end calibration has the effect of adapting the computation to the shapes of the metalens PSFs, which differ substantially from Gaussians.
To analyze the depth accuracy, we measured the depths of test objects at a series of known distances and compared them with the true object distances. The test objects were textured planes oriented parallel to the lens plane. At each object distance, the mean deviation of depth, , was computed by using pixels that have confidence values greater than a threshold. Fig. 4B shows the measured depth for different confidence thresholds as a function of object distance. For a confidence threshold of 0.5, the measured depth was accurate to within a mean deviation below or around of the true depth, over a range of true object distances between 0.3 and 0.4 m. Beyond this range, the measured depth defaulted to the extreme depth value that the system can predict, as indicated by plateaus on the left and right ends. This indicated that the 2 images were so blurry that there was an insufficient contrast difference between them.
Fig. 5 shows depth maps for a variety of scenes. Because it uses a single shot, the metalens depth sensor can measure objects that move, such as the fruit flies and water stream in Fig. 5 A and B. It can also measure the depth of translucent entities, such as the candle flames of Fig. 5C, that cannot typically be measured by using active sensors like Lidar and time-of-flight. Fig. 5D shows a slanted plane with printed text, where the blur change between the 2 images is particularly apparent. In general, the sensor reports a larger number of depth measurements near regions with edges and texture, whereas regions with uniform intensity and low contrast are typically discarded as having low confidence values. Note that the blur differences between and are visually apparent in Fig. 5 A, B, and D, but that the system still succeeds in Fig. 5C, where the differences in blur are hard to discern.
Fig. 5.

For scenes other than Fig. 5C, we used green LED light sources, and the overall transmission efficiency of the metalens plus the bandpass filter was around . For sunlight illumination, the bandpass filter transmitted around of the visible light of the solar spectrum. The absolute irradiance that supports the function of the sensor varied based on the sensitivity of the photosensor that was used and can be estimated from specifications including absolute sensitivity threshold, dynamic range, etc. For our experimental setup, the irradiance at the aperture was estimated to be between 0.3 and 0.5 W/m2 within the working bandwidth to support the function of the sensor.
The sensor generates depth and confidence maps of 400400 pixels at more than 100 frames per second using a combined central processing unit and graphics processing unit (Intel i5 8500k and NVIDIA TITAN V). It could be accelerated substantially by optimizing the code and/or the hardware because the calculations are spatially localized and few in number. Producing the depth and confidence values at each output pixel required 637 FLOPs and involved only the 25 25 surrounding pixels. For context, an efficient implementation of a binocular stereo algorithm requires about 7,000 FLOPs per output pixel (37), and a system-on-chip implementation of the well-known Lucas–Kanade optical flow algorithm (with spatial dependence similar to that of our sensor) requires over 2,500 FLOPs per pixel (38).
Discussion
The metalens depth sensor inherits some of the limitations that exist in the vision system of jumping spiders, such as a limited spectral bandwidth and a limited field of view. However, these limits are not fundamental, and they can be alleviated by more sophisticated metalens designs. The spectral bandwidth can be expanded by using achromatic metalenses (35, 39, 40), which also improve light efficiency. The field of view can be improved, for example, by using metalens nanopillars that are sensitive to polarization to induce 2 differently focused images that are superimposed on the sensor plane with orthogonal polarizations (36) and then transducing the 2 images with a spatially multiplexed, polarization-sensitive sensor array. This would effectively trade spatial resolution and light efficiency for an increase in field of view.
The proposed computational algorithm produces a dense field of depth estimates that are each associated with a confidence value. The confidence is essential for the users of the depth sensor to remove unreliable predictions. It also uses a multiscale filtering approach to handle image textures at different spatial frequency and takes advantage of the confidence to merge all different spatial scales together, compared to previous methods (20, 22) that only use filters at a single, predetermined spatial scale. The proposed algorithm does not incorporate inference-based methods such as Markov random fields (MRFs) or conditional random fields (CRFs) that could exploit longer-range coherence between depth values across the field of view. Instead, the depth and confidence estimations at each pixel are only based on information of its spatial neighborhood. The advantages of this design choice are flexibility and generality. For tasks that require high speeds, the output can be used as-is, with simple thresholding of confidence values. For tasks that require higher accuracy and fewer holes in the depth map, the current output can be fed into an MRF/CRF (or any other spatial regularizer) that is appropriate for that task. Moreover, because the pipeline is end-to-end differentiable, its parameters can be fine-tuned in conjunction with MRF/CRF parameters to optimize performance on the specific task.
By combining cutting-edge nanotechnology and computer vision algorithms, this work introduces a passive snapshot depth sensor that mimics some of the capabilities of a jumping spider. The sensor’s small volume, weight, and computation (i.e., power) bring depth-sensing capabilities closer to being feasible on insect-scale platforms, such as microrobots, ingestible devices, far-flung sensor networks, and small wearable devices. Combinations of nanophotonics and efficient computation that are different from the ones in this paper might lead to other forms of compact visual sensors, and this is an area that remains relatively unexplored (41).
Acknowledgments
This project was supported by Air Force Office of Scientific Research Multidisciplinary University Research Initiative Grants FA9550-14-1-0389 and FA9550-16-1-0156; and NSF Award IIS-1718012. Y.-W.H. and C.-W.Q. are supported by the National Research Foundation, Prime Minister’s Office, Singapore under Competitive Research Program Award NRF-CRP15-2015-03. E.A. is supported by NSF Graduate Research Fellowship DGE1144152. Metalens fabrication was performed at Harvard’s Center for Nanoscale Systems, supported by NSF Grant 1541959. Q.G. and Z.S. thank Mohammadreza Khorasaninejad for helpful discussions. Z.S. thanks Zhehao Dai for helpful comments and discussions.
Supporting Information
Appendix (PDF)
- Download
- 2.14 MB
Movie S1.
The video shows real time depth estimation using the metalens depth sensor for several dynamic scenes. Similar to Fig. S17, it simultaneously shows the captured image pairs (I+,I−), and the measured depth map Z masked using the confidence metric.
- Download
- 34.77 MB
References
1
P. F McManamon, Field Guide to Lidar (SPIE Press, Bellingham, WA, 2015).
2
S. Achar, J. R Bartels, W. L. Whittaker, K. N. Kutulakos, S. G Narasimhan, Epipolar time-of-flight imaging. ACM Trans. Graph. 36, 37 (2017).
3
M. Gupta, S. K Nayar, M. B Hullin, J. Martin, Phasor imaging: A generalization of correlation-based time-of-flight imaging. ACM Trans. Graph. 34, 156 (2015).
4
M. Hansard, S. Lee, O. Choi, R. P. Horaud, Time-of-Flight Cameras: Principles, Methods and Applications (Springer Science & Business Media, New York, 2012).
5
F. Heide, W. Heidrich, M. Hullin, G. Wetzstein, Doppler time-of-flight imaging. ACM Trans. Graph. 34, 36 (2015).
6
S. Shrestha, F. Heide, W. Heidrich, G. Wetzstein, Computational imaging with multi-camera time-of-flight systems. ACM Trans. Graph. 35, 33 (2016).
7
S. Su, F. Heide, G. Wetzstein, W. Heidrich, “Deep end-to-end time-of-flight imaging” in Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), M. Brown, B. Morse, S. Peleg, Eds. (IEEE Computer Society, Los Alamitos, CA, 2018), pp. 6383–6392.
8
R. Xu, S. K Nayar, M. Gupta, “Trapping light for time of flight” in Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), M. Brown, B. Morse, S. Peleg, Eds. (IEEE Computer Society, Los Alamitos, CA, 2018), pp. 6219–6227.
9
J. Geng, Structured-light 3D surface imaging: A tutorial. Adv. Opt. Photon 3, 128–160 (2011).
10
P. Mirdehghan, W. Chen, K. N Kutulakos, “Optimal structured light a la carte” in Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), M. Brown, B. Morse, S. Peleg, Eds. (IEEE Computer Society, Los Alamitos, CA, 2018), pp. 6248–6257.
11
N. Lazaros, G. C. Sirakoulis, A. Gasteratos, Review of stereo vision algorithms: From software to hardware. Int. J. Optomechatronics 2, 435–462 (2008).
12
R. Ng et al., Light field photography with a hand-held plenoptic camera (Comput. Sci. Tech. Rep. 2005-02, Stanford University, Stanford, CA, 2005).
13
R. J. Lin et al., Achromatic metalens array for full-colour light-field imaging. Nat. Nanotechnol. 14, 227–231 (2019).
14
H. Lin, C. Chen, S. Bing Kang, J. Yu, “Depth recovery from light field using focal stack symmetry” in Proceedings of the 2015 IEEE International Conference on Computer Vision, H. Bischof, D. Forsyth, C. Schmid, S. Sclaroff, Eds. (IEEE Computer Society, Los Alamitos, CA, 2015), pp. 3451–3459.
15
K. Y Ma, S. M Felton, R. J Wood, “Design, fabrication, and modeling of the split actuator microrobotic bee” in Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), A. T. de Almeida, U. Nunes, Eds. (IEEE Computer Society, Los Alamitos, CA, 2012), pp. 1133–1140.
16
M. Rubenstein, A. Cornejo, R. Nagpal, Programmable self-assembly in a thousand-robot swarm. Science 345, 795–799 (2014).
17
E. Y Erdem et al., Thermally actuated omnidirectional walking microrobot. J. Microelectromech. Syst. 19, 433–442 (2010).
18
E. Alexander, Q. Guo, S. Koppal, S. J. Gortler, T. Zickler, Focal flow: Velocity and depth from differential defocus through motion. Int. J. Comput. Vis. 126, 1062–1083 (2018).
19
Q. Guo, E. Alexander, T. Zickler, “Focal track: Depth and accommodation with oscillating lens deformation” in Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), K. Ikeuchi, G. Medioni, M. Pelillo, Eds. (IEEE Computer Society, Los Alamitos, CA, 2017), pp. 966–974.
20
M. Subbarao, G. Surya, Depth from defocus: A spatial domain approach. Int. J. Comput. Vis. 13, 271–294 (1994).
21
A. P. Pentland, A new sense for depth of field. IEEE Trans. Pattern Anal. Mach. Intell. PAMI-9, 523–531 (1987).
22
M. Watanabe, S. K. Nayar, Rational filters for passive depth from defocus. Int. J. Comput. Vis. 27, 203–225 (1998).
23
H. Farid, E. P. Simoncelli, Range estimation by optical differentiation. JOSA A 15, 1777–1786 (1998).
24
G. Menda, P. S. Shamble, E. I Nitzany, J. R Golden, R. R. Hoy, Visual perception in the brain of a jumping spider. Curr. Biol. 24, 2580–2585 (2014).
25
M. F. Land, Structure of the retinae of the principal eyes of jumping spiders (Salticidae: Dendryphantinae) in relation to visual optics. J. Exp. Biol. 51, 443–470 (1969).
26
T. Nagata et al., Depth perception from image defocus in a jumping spider. Science, 335, 469–471 (2012).
27
N. F. Yu, F. Capasso, Flat optics with designer metasurfaces. Nat. Mater. 13, 139–150 (2014).
28
M. Khorasaninejad et al., Multispectral chiral imaging with a metalens. Nano Lett. 16, 4595–4600 (2016).
29
M. Khorasaninejad et al., Metalenses at visible wavelengths: Diffraction-limited focusing and subwavelength resolution imaging. Science 352, 1190–1194 (2016).
30
A. V. Kildishev, A. Boltasseva, V. M. Shalaev, Planar photonics with metasurfaces. Science 339, 1232009 (2013).
31
A. Silva et al., Performing mathematical operations with metamaterials. Science 343, 160–163 (2014).
32
S. Yi et al., Subwavelength angle-sensing photodetectors inspired by directional hearing in small animals. Nat. Nanotechnol. 13, 1143–1147 (2018).
33
D. M. Lin, P. Y. Fan, E. Hasman, M. L. Brongersma, Dielectric gradient metasurface optical elements. Science 345, 298–302 (2014).
34
R. C. Devlin, M. Khorasaninejad, W. T. Chen, J. Oh, F. Capasso, Broadband high-efficiency dielectric metasurfaces for the visible spectrum. Proc. Natl. Acad. Sci. U.S.A. 113, 10473–10478 (2016).
35
W. T. Chen et al., A broadband achromatic metalens for focusing and imaging in the visible. Nat. Nanotechnol. 13, 220–226 (2018).
36
J. P. B. Mueller, N. A. Rubin, R. C. Devlin, B. Groever, F. Capasso, Metasurface polarization optics: Independent phase control of arbitrary orthogonal states of polarization. Phys. Rev. Lett. 118, 113901 (2017).
37
S. Rotheneder, “Performance analysis of a stereo matching implementation in OpenCL,” Thesis, Technische Universität Wien, Vienna (2018).
38
D. Bagni, P. Kannan, S. Neuendorffer, Demystifying the Lucas-Kanade optical flow algorithm with Vivado HLS (Tech. note XAPP1300, Xilinx, San Jose, CA, 2017).
39
Z. Shi et al., Single-layer metasurface with controllable multiwavelength functions. Nano Lett. 18, 2420–2427 (2018).
40
M. Khorasaninejad et al., Achromatic metalens over 60 nm bandwidth in the visible and metalens with reverse chromatic dispersion. Nano Lett. 17, 1819–1824 (2017).
41
D. G. Stork, P. R. Gill, “Special-purpose optics to reduce power dissipation in computational sensing and imaging systems” in Proceedings of 2017 IEEE SENSORS, D. Uttamchandani, K. Ozanyan, Eds. (IEEE, Piscataway, NJ, 2017), pp. 1197–1199.
Information & Authors
Information
Published in
Classifications
Copyright
Copyright © 2019 the Author(s). Published by PNAS. This open access article is distributed under Creative Commons Attribution License 4.0 (CC BY).
Submission history
Published online: October 28, 2019
Published in issue: November 12, 2019
Keywords
Acknowledgments
This project was supported by Air Force Office of Scientific Research Multidisciplinary University Research Initiative Grants FA9550-14-1-0389 and FA9550-16-1-0156; and NSF Award IIS-1718012. Y.-W.H. and C.-W.Q. are supported by the National Research Foundation, Prime Minister’s Office, Singapore under Competitive Research Program Award NRF-CRP15-2015-03. E.A. is supported by NSF Graduate Research Fellowship DGE1144152. Metalens fabrication was performed at Harvard’s Center for Nanoscale Systems, supported by NSF Grant 1541959. Q.G. and Z.S. thank Mohammadreza Khorasaninejad for helpful discussions. Z.S. thanks Zhehao Dai for helpful comments and discussions.
Authors
Competing Interests
Competing interest statement: F.C. is a cofounder of Metalenz. D.S. is on the Scientific Advisory Board of Metalenz.
Metrics & Citations
Metrics
Altmetrics
Citations
Cite this article
Compact single-shot metalens depth sensors inspired by eyes of jumping spiders, Proc. Natl. Acad. Sci. U.S.A.
116 (46) 22959-22965,
https://doi.org/10.1073/pnas.1912154116
(2019).
Copied!
Copying failed.
Export the article citation data by selecting a format from the list below and clicking Export.
Cited by
Loading...
View Options
View options
PDF format
Download this article as a PDF file
DOWNLOAD PDFLogin options
Check if you have access through your login credentials or your institution to get full access on this article.
Personal login Institutional LoginRecommend to a librarian
Recommend PNAS to a LibrarianPurchase options
Purchase this article to access the full text.