SPIE Digital Library Get updates from SPIE Newsroom
  • Newsroom Home
  • Astronomy
  • Biomedical Optics & Medical Imaging
  • Defense & Security
  • Electronic Imaging & Signal Processing
  • Illumination & Displays
  • Lasers & Sources
  • Micro/Nano Lithography
  • Nanotechnology
  • Optical Design & Engineering
  • Optoelectronics & Communications
  • Remote Sensing
  • Sensing & Measurement
  • Solar & Alternative Energy
  • Sign up for Newsroom E-Alerts
  • Information for:
    Advertisers
SPIE Photonics West 2017 | Register Today

SPIE Defense + Commercial Sensing 2017 | Call for Papers

Journal of Medical Imaging | Learn more

SPIE PRESS




Print PageEmail PageView PDF

Electronic Imaging & Signal Processing

Superior vision in nocturnal insects inspires new night vision technologies

Algorithms that dramatically improve the quality of video sequences captured in very dim light have been developed on the basis of the neural mechanisms in nocturnal insects with excellent visual capabilities.
11 April 2016, SPIE Newsroom. DOI: 10.1117/2.1201603.006426

At night—because of the paucity of light—human visual capacities are severely reduced, with a complete loss of the ability to see color and a dramatic loss in the ability to see fine spatial and temporal details. However, this is not the case for many nocturnal animals (e.g., insects). Nocturnal animals typically have highly sensitive eye designs and visual neural circuitry that is optimized for the extraction of reliable information from dim and noisy visual images. Although we are only now at the threshold of understanding the neural mechanisms responsible for reliable nocturnal vision, there is growing evidence to suggest that the neural summation of photons in space and time is critically important.1, 2 Artificial imaging systems that operate in dim light—like nocturnal eyes—also suffer from limitations in detection reliability. These systems should thus also benefit from the same strategies for improving their reliability (i.e., the summation of photons in space and time).

Purchase SPIE Field Guide to Image ProcessingIn our recent work—particularly on fast-flying moths and bees, and on ball-rolling dung beetles—we have studied the visual abilities of nocturnal animals. In particular, we have shown that they are able to distinguish colors3—see Figure 1(A)—detect faint movements,1 learn visual landmarks4—see Figure 1(B)—orient to the faint pattern of polarized light produced by the Moon,5 and navigate using the stars.6 These impressive visual capabilities are the result of exquisitely adapted eyes and visual systems (the product of millions of years of evolution).


Figure 1. Nocturnal insects with sensitive vision. (A) The European elephant hawkmoth, Deilephila elpenor, whose sensitive superposition eyes allow it to distinguish color at night. (B) The central American halictid bee, Megalopta genalis, whose sensitive apposition eyes allow it to forage at night (by visually learning landmarks along the foraging route and around the nest entrance). Photographs provided with kind permission of Michael Pfaff (A) and Ajay Narendra (B).

The strategy for improving nocturnal vision involves improving sensitivity by increasing the size of the visual ‘pixel’ and by lengthening the visual exposure time. Even though vision in dim light thereby becomes necessarily coarser and slower, it also becomes significantly more reliable.1, 2 In our most recent work,7 it was therefore our idea to take visual solutions—which result from millions of years of biological evolution—and incorporate them in a computer algorithm that radically improves the quality of video sequences collected in dim light. Our algorithm thus mimics several of the key steps of nocturnal visual processing.8

The first step in our algorithm (see Figure 2) involves an amplification of the dim image sequence. Such amplification (or gain) is also typical of the photoreceptor responses of nocturnal insects.9 The downside of this amplification is that it amplifies both the signal and the noise, and thus does not improve the signal-to-noise ratio per se. To negotiate this issue, we include another crucial step in the algorithm that involves spatial and temporal summation. This step results in signal information (which is correlated between frames in the video sequence) being summed and reinforced, and noise (which is uncorrelated) being averaged out. The result is a dramatic improvement in the signal-to-noise ratio. Moreover, because this summation is conducted separately (and in parallel) for the red, green, and blue channels, the summation result preserves color information (as evident in the output video frame in Figure 2). A similar, separate and parallel, summation of color channels in nocturnal animals is likely the reason that many can see color at night.


Figure 2. The main steps in the night vision algorithm. The algorithm consists of a contrast enhancement step (amplification) and a noise-reduction step that is based on optimal spatial and temporal summation. The algorithm adapts automatically to the local intensity distribution and preserves as much of the structure in the scene as possible. It also adapts to the local noise level, which depends on the ambient light level. Furthermore, it sharpens edges in the output at appropriate places to prevent over-smoothing (via a lateral inhibition algorithm that is similar to lateral inhibition in the retina). A still frame (upper left) from a dim input video sequence can be compared with a still frame (upper right) from the significantly enhanced output video sequence, i.e., which resulted from processing with the algorithm.

A disadvantage of the even radial spatial summation, however, is that it does not preserve edges and boundaries (i.e., the contrast elements of visual scenes that define and localize objects). To deal with this problem, our algorithm performs a real-time structure tensor analysis of the intensity structure (in space and time) of the scene contained within the film. In addition, local smoothing kernels for each location in the film frame are constructed. These kernels adapt to the local spatiotemporal intensity structure. They are large (i.e., summing more) in directions where there are no spatiotemporal edges, and small (i.e., summing less) in directions where there are strong edges. Moreover, because dimmer video sequences require greater extents of summation, the volumes of these kernels become large for videos filmed at lower light levels. The algorithm therefore adapts to light intensity in an automatic manner, much like an animal's visual system.

By mimicking the way nocturnal visual systems process information, our night-vision algorithm can be used to dramatically restore the quality of video sequences recorded in very dim light.8 Not only are spatial and temporal details—obscured by noise in the original video—restored, but color information is also preserved. Our hypothesis that spatial and temporal summation is a crucial strategy in nocturnal visual processing is thus very well supported by our algorithm. The output of the algorithm suggests that summation can greatly enhance nocturnal visual performance.

Insects have conquered the terrestrial world's dimmest habitats and have done so with some of the most sensitive visual systems found in the animal kingdom. In our work we have shown that the neural strategies they use—and the benefits to visual reliability they bestow—are directly applicable to algorithms for processing video sequences captured in very dim light. As we continue to study how the brain of nocturnal insects processes and optimizes the visual information it receives, we will hopefully be in a position to further improve our algorithm. We plan to add components that are necessary for tasks such as object recognition and distance discrimination.

I am grateful to many institutions that have supported this work over the years, including the Swedish Research Council, the US Air Force Office of Scientific Research, the Royal Physiographic Society of Lund, the Crafoord Foundation, the Wenner-Gren Foundation, the Swedish International Development Agency, Toyota Research and Development, and the University of Lund.


Eric Warrant
University of Lund
Lund, Sweden

Eric Warrant received his PhD from the Australian National University in 1990. He studies vision and visual navigation in animals from extremely dim habitats (nocturnal and deep sea). His research has led to the discovery of neural principles that permit vision in dim light.


References:
1. A. L. Stöckl, D. C. O'Carroll, E. J. Warrant, Neural summation in the hawkmoth visual system extends the limits of vision in dim light, Curr. Biol. 26, p. 821-826, 2016. doi:10.1016/j.cub.2016.01.030
2. E. J. Warrant, Seeing better at night: life style, eye design, and the optimum strategy of spatial and temporal summation, Vision Res. 39, p. 1611-1630, 1999.
3. A. Kelber, A. Balkenius, E. J. Warrant, Scotopic colour vision in nocturnal hawkmoths, Nature 419, p. 922-925, 2002.
4. E. J. Warrant, A. Kelber, A. Gislén, B. Greiner, W. Ribi, W. T. Wcislo, Nocturnal vision and landmark orientation in a tropical halictid bee, Curr. Biol. 14, p. 1309-1318, 2004.
5. M. Dacke, D.-E. Nilsson, C. H. Scholtz, M. Byrne, E. J. Warrant, Insect orientation to polarized moonlight, Nature 424, p. 33, 2003.
6. M. Dacke, E. Baird, M. Byrne, C. H. Scholtz, E. J. Warrant, Dung beetles use the Milky Way for orientation, Curr. Biol. 23, p. 298-300, 2013.
7. E. J. Warrant, Superior visual performance in nocturnal insects: neural principles and bio-inspired technologies, Proc. SPIE 9797, p. 979703, 2016. doi:10.1117/12.2218336
8. E. J. Warrant, M. Oskarsson, H. Malm, The remarkable visual abilities of nocturnal insects: neural principles and bioinspired night-vision algorithms, Proc. IEEE 102, p. 1411-1426, 2014.
9. R. Frederiksen, W. T. Wcislo, E. J. Warrant, Visual reliability and information rate in the retina of a nocturnal bee, Curr. Biol. 18, p. 349-353, 2008.