Next Article in Journal
Ground Contact Force and Moment Estimation for Human–Exoskeleton Systems Using Dynamic Decoupled Coordinate System and Minimum Energy Hypothesis
Next Article in Special Issue
Colorful 3D Reconstruction and an Extended Depth of Field for a Monocular Biological Microscope Using an Electrically Tunable Lens
Previous Article in Journal
Maximization of Power Density of Direct Methanol Fuel Cell for Greener Energy Generation Using Beetle Antennae Search Algorithm and Fuzzy Modeling
Previous Article in Special Issue
Advances in Ghost Imaging of Moving Targets: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Advanced Biomimetic Multispectral Curved Compound Eye Camera for Aerial Multispectral Imaging in a Large Field of View

1
Key Laboratory of Spectral Imaging Technology, Xi’an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, No. 17, Xinxi Road, Xi’an 710119, China
2
Center of Materials Science and Optoelectronics Engineering, University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Biomimetics 2023, 8(7), 556; https://doi.org/10.3390/biomimetics8070556
Submission received: 27 September 2023 / Revised: 14 November 2023 / Accepted: 15 November 2023 / Published: 20 November 2023
(This article belongs to the Special Issue Bionic Imaging and Optical Devices)

Abstract

:
In this work, we demonstrated a new type of biomimetic multispectral curved compound eye camera (BM3C) inspired by insect compound eyes for aerial multispectral imaging in a large field of view. The proposed system exhibits a maximum field of view (FOV) of 120 degrees and seven-waveband multispectral images ranging from visible to near-infrared wavelengths. Pinhole imaging theory and the image registration method from feature detection are used to reconstruct the multispectral 3D data cube. An airborne imaging experiment is performed by assembling the BM3C on an unmanned aerial vehicle (UAV). As a result, radiation intensity curves of several objects are successfully obtained, and a land type classification is performed using the K-means method based on the aerial image as well. The developed BM3C is proven to have the capability for large FOV aerial multispectral imaging and shows great potential applications for distant detecting based on aerial imaging.

1. Introduction

Unmanned aerial vehicles (UAV) are commonly used in multispectral remote imaging, with the advantages of a low price, easy use and flexible mobility [1]. Among the multiple kinds of UAVs, quadrotor and six-rotor UAVs are widely applied in aerial platforms for multispectral snap-shot imaging. In general, these UAVs have a payload capacity under 7 kg, a speed of 12 m/s and a flight ceiling altitude of 500 m [2]. There are different kinds of products for airborne multispectral imaging, like ADC lite (TetracamInc., Chatsworth, CA, USA), RedEdge (Micasensen Inc., Seattle, WA, USA), Parrot Sequoia (Parrot Inc., Paris, France), etc.
UAV-mounted multispectral cameras are supposed to be small and light and have a large FOV, which fits well with the features of artificial compound eye (ACE) systems. ACE is a kind of biomimetic imaging system inspired by the compound eye structure of insects. The compound eye in nature contains multiple ommatidia arranged on a free-shaped eyeball, and each of the ommatidia has its own photoreceptor. The structure allows for imaging independently and achieves a large FOV with a compact size.
The ommatidia of ACE are usually designed as separate imaging channels, whose optical system is optimized according to the software for aberration correction. By now, many research works about ACE systems have been reported in order to take advantage of its features. In general, compound eye imaging systems include planar and curved compound eye systems. Based on the Thin Observation Module by Bound Optics (TOMBO) system proposed in 2000 [3], the planar compound eye system contains two major parts: a planar lens array and an image sensor. These systems achieve high-resolution, high-speed imaging and 3D information acquisition [4,5,6,7]. The curved compound eye system contains a curved lens array with a planar or curved image sensor, whose structure is closer to the insect compound eyes in nature. In the research, the curved compound eye system shows the advantage of ultra-large FOV imaging with low distortion. The application of curved systems mainly focuses on shape detecting, trajectory tracking, orientation detecting and 3D measurement [8,9,10,11,12,13].
As each ommatidium works separately in a compound eye, the insects have been found to specialize different parts of the compound eye for different utilities [14,15]. This discovery inspires the development of multispectral ACE systems. Most of the multispectral ACE systems are based on the TOMBO system and the color TOMBO system in 2003 [16]. Following the work of Tanida’s group, Shogenji et al. proposed a multispectral imaging system based on the TOMBO structure by integrating narrow-band filters to ommatidia lenses [17]. In 2008, Scott A. Mathews et al. proposed a developed TOMBO multispectral imaging system with 18 lens imaging units with filters [18]. In 2010, Kagawa et al. reported a multispectral imaging system on the basis of the TOMBO structure that utilizes the rolling shutter CMOS to achieve high-speed multispectral imaging [19]. Other similar multispectral imaging systems are widely proposed for different applications [20,21,22].
The multispectral ACE system utilized on the UAV platform was first achieved in 2020 by Nakanshi et al. from Osaka University [23]. In the article, a prototype with an embedded computer and a multispectral TOMBO imaging system was demonstrated as shown in Figure 1. Multispectral images of a ground oil furnace and vehicle target were captured with the system at a flight altitude of 3–7 m and were also analyzed with the normalized vegetation index (NDVI). This work shows the capacity of the multispectral ACE system for aerial imaging.
In previous studies, we proposed a so-called Multispectral Curved Compound Eye Camera (MCCEC) in 2020 [24]. The system was designed to be able to achieve an FOV of 120 degrees with seven-wavebands multispectral imaging, with a rather short focal length of 0.4 mm. Based on the design, a biomimetic multispectral curved compound eye camera (BMCCEC) prototype was demonstrated in 2021. The BMCCEC system achieved a maximum FOV of 98 degrees and an improved focal length of 5 mm [25]. However, the structure design of the prototype causes a blind area and lacks mounting space for filters, which is limited in the prototype aerial imaging.
To solve these issues of the BMCCEC system and explore the possibility of curved ACE systems in aerial imaging, an advanced BMCCEC (BM3C) system is proposed in this article. Based on design theory analysis, the system achieves multispectral imaging without a blind area, with a maximum FOV of 127.4 degrees, a focal length of 2.76 mm and seven multispectral channels. To test the performance of the system, spectral calibration and correction, an airborne imaging experiment and data processing are performed with the prototype.

2. Multispectral Imaging System

2.1. System Design Theory

The BM3C system achieves multispectral imaging based on clusters of multispectral ommatidia. Thus, the arrangement of the ommatidia, FOV and focal length are solved theoretically to avoid a blind area and image overlap in this work.

2.1.1. The Arrangement of the Multispectral Lens Array

ACE systems contain multiple methods of lens arrangement. To achieve multispectral imaging, the included angle between the neighbor channels should be as consistent as possible to avoid wasting the FOV. Based on the tessellation rule, the most common arrangements are regular triangles and quadrilaterals. Three possible arrangements are shown in Figure 2. It can be seen that the arrangement based on regular triangles has a consistent distance between nodes, as well as the maximum number of spectrum channels without FOV waste.
Thus, an arrangement method based on triangle mesh is proposed, as shown in Figure 3a. As the highlighted one shows, each cluster unit contains seven spectral channels. The mesh is then projected to a curved shell to determine the actual ommatidia arrangement, as shown in Figure 3b.
To determine the central wavelengths of seven channels, a survey of commercial aerial multispectral imaging systems is made. Representative products contain imaging channels of blue, green, red and NIR wavebands for agriculture and ground object classification applications. Therefore, the working waveband of the system is chosen as 440 nm–800 nm. For the prototype system, seven central wavelengths are designed as 500, 560, 600, 650, 700, 750 and 800 nm to acquire the critical spectral information of ground objects with green leaves, and this working waveband can cover the so called ‘red edge’ reflection spectrum of the green leaves [26].

2.1.2. The FOV

In the design of BM3C, 169 ommatidia are arranged on the sphere structure with the sphere projection method [25]. While the included angle between the light axes of two neighbor ommatidia channels is seven degrees, the system should achieve a maximum FOV of 120 degrees and a maximum multispectral imaging field of 98 degrees.
To avoid a blind area, the FOV of each ommatidium is calculated with an overlap calculation equation [27], as shown in Figure 4.
d c = d sin β 2 tan α = [ R cos β 2 + R sin β 2 cos ( α     β 2 ) / sin ( α     β 2 ) ] sin β 2 tan α
where the radius of the curved shell R is 125 mm, α represents the half FOV of each channel and β represents the included angle between the light axes of a pair of neighbored multispectral channels of the same waveband. By solving the included angle in the 3D model of ommatidia distribution, the maximum angle β is 18.52 degrees according to the result. Considering the working environment of the airborne camera, the minimum working distance is set as 1000 mm. Under this condition, the minimum FOV of the ommatidium is solved as ±10.57 degrees. The FOV is set as ±12 degrees during the design procedure to leave space for machining and adjustment allowance.

2.1.3. The Focal Length

To avoid the overlap between neighboring sub-images, there should be enough of a physical interval between the neighbor ommatidia. According to a previous study, there should be a minimum interval of 0.4 mm on the imaging plane. Thus, the focal length f of the whole optical system can be solved with Equation (2).
r = s p   ×   n p     n f   ×   l f 2   ×   n o f = r tan θ
where s p and n p represent the size and number of the pixels of the CMOS sensor. n f = 14 and l f = 0.4 mm represent the number and length of the interval between the image of neighbor ommatidia. n o = 15 is the number of the ommatidia on the diagonal of the hexagon array. θ = 12 degrees is half of the FOV of each ommatidium channel. From Equation (2), the focal length f of the system is solved as 2.7 mm. The focal length f is distributed into the two optical subsystems with the Newton formula.

2.2. Prototype of BM3C

As the critical parameters are solved with theory analysis, a prototype of BM3C is designed. In the optical design process, the system is optimized with optical design software. Figure 5 shows the 3D model of the system, as well as two critical parameters from the simulation, the Modulation Transfer Function (MTF) and the distortion of the edge channel of the system. The simulation system has an MTF larger than 0.36 at 55 lp/mm and a distortion less than 2% in all channels, which represents a fine imaging quality.
According to the design result, a camera prototype is manufactured, as shown in Figure 6, which contains three parts.
  • A curved shell that contains 169 multispectral ommatidia; each ommatidium is a doublet lens with a focal length of 17.02 mm. Seven groups of narrowband filters are mounted before the ommatidia. The whole shell contains 25 filters with a nominal central wavelength of 500 nm and 6 groups of 24 filters with nominal central wavelengths at 560, 600, 650, 700, 750 and 800 nm; the arrangement of the filters is shown in Figure 2.
  • An optical relay system with eight glass lenses, which transforms the curved image plane of the ommatidia array into a planar one. According to the Newton formula, the focal length is designed as 1 mm to achieve a whole system focal length of 2.7 mm.
  • An Imperx Cheetah C5180M CMOS camera as the image sensor, with a resolution of 5120 × 5120 and a pixel size of 4.5 μm.
Several performance tests are conducted based on the BM3C system. The important parameters of the camera are listed in Table 1.

3. Image Reconstruction Method

3.1. Method Based on Reprojection and Feature Detection

3.1.1. Error Analysis of the Image Reprojection

In the former study [25], a calibration-free image reconstruction method was proposed, as shown in Figure 7a. The method may introduce errors to the image reconstruction procedure in two major variables, the x and y coordinates of the principal point and the focal length of the system. Figure 7b shows an example of the error introduction. The reprojection method causes a large image registration error and leads to inaccurate multispectral measurement results.
The error introduced by the image reconstruction method can be described with Equation (3).
Δ yo = Δ o ( d   +   f ) f Δ yf = Δ f dx f   ( f     Δ f )
where f represents the theoretical focal length of the channel, O represents the theoretical coordinate of the principal point, Δ f and Δ O represent the error of the focal length and principal point and Δ yf and Δ yo represent the errors caused by the introduced errors of two variables. According to Equation (3), the reconstruction error caused by the principal point only relates to Δ O , and the reconstruction error caused by the focal length has a positive correlation with the distance x between the image point and principal point. Thus, to reduce the reconstruction error, an image reconstruction method combined with feature detection and image registration is proposed.

3.1.2. The Combined Reconstruction Method

The image reconstruction method may lead to an error from the position of the image pixel. A new reconstruction method combines feature detection and the reprojection method to reduce the error of image reconstruction. This combined image reconstruction method is shown in Figure 8 and contains three main procedures:
  • First, the positions of all sub-images of ommatidia are found on the image plane with a generated and adjusted hexagon grid. As the center and radius of each circle image are determined, 127 multispectral clusters are noted by looking at the list of ommatidia and searching for the six nearest neighbor ommatidia for each ommatidium, except for the ones on the edge.
  • After the multispectral clusters are noted, the image registration is performed based on the SIFT (Scale-Invariant Feature Transform) feature extraction algorithm, and the homography matrixes of the six surrounding channels in relation to the main channel are acquired with feature matching and the mismatching is reduced with the RANSAC (RANdom SAmple Consensus) algorithm [28]. The thresholds of the algorithm are manually adjusted to achieve the best matching and registration results.
  • With the homography matrix and the image registration method, the projected positions of each pixel in the valid imaging area in the central ommatidium of each cluster are calculated and noted in a look-up table to achieve real-time multispectral image reconstruction.
The reconstructed image of a multispectral cluster unit is shown in Figure 9a, and Figure 9b shows the overlap map of the cluster. In Figure 9b, the white part in the center represents the valid imaging area of seven wavebands, and the surrounding areas represent the areas where fewer than seven wavebands overlap.
Compared with the reprojection method, the combined method achieves a lower registration error of fewer than two pixels at the same place. Figure 10 shows a feature point in the image and two spectral channels with reconstruction deviation, and the reprojection method shows a larger deviation. Thus, the combined reconstruction method leads to a multispectral image acquisition with a higher precision, and the look-up table makes the program more than four times faster.

3.1.3. Working Distance

As a compound eye system, the baseline and included angle between the light axis of ommatidia generate optical parallax. On one hand, this provides the capability to achieve distance measurement and 3D information acquisition; on the other hand, it causes image mismatching while using a look-up table in a short distance. Thus, the relationship between the parallax and the object distance has to be evaluated to determine the working distance of the system.
The imaging parallax of a pair of neighbor ommatidia is shown in Figure 11:
which can be solved with Equation (4):
b = R tan θ c = d     R cos θ   +   R a = b 2   +   c 2     2 b c cos ( 90   +   θ ) x = f tan { 90     arcsin [ c sin ( 90   +   θ ) a ] } r e s = x   +   x
where the imaging procedure of an ommatidium channel is simplified as a pinhole camera model. θ is the included angle between the light axis OA and the object position OB, while the neighbor ommatidium has an included angle of 7 − θ. Both angles are 3.5 degrees for simplification. d is the distance of the object point. f = 2.7 mm is the focal length of the optical system. And R = 125 mm is the radius of the entrance pupil sphere. The parallax Res is a function relating to the distance d.
To avoid the mismatching caused by the parallax, the relationship between the object distance and parallax is solved, as shown in Figure 12.
The variation in the parallax caused by the system remains fewer than 0.1 pixels when the object distance is large enough, which can be assumed to be invariant in image reconstruction. Thus, the parallax with an object distance of 100 m is chosen as the static value, and the object distance of the pre-matching target should be larger than 8.4 m.

3.2. Method of Multispectral Information Acquisition

3.2.1. Radiometric Correction Based on Calibration

As described, the multispectral information of a single target is reconstructed from the image captured by a cluster of multispectral ommatidia. To invert the original radiation characteristics of the objects, spectral calibration and radiometric calibration are performed with the BM3C.
The spectral calibration applies the monochromator as the light source to measure the central wavelength and full-width-at-half-maximum (FWHM) of each spectrum channel. The monochromator works in the scanning mode, the wavelength scan range is 475–825 nm and the step is 2 nm. Figure 13 shows a set of images captured by an ommatidium during scanning. By extracting the Digital Number (DN) value of each pixel and a fitting procedure with a gaussian function, the central wavelength and half-band width of the spectrum channels are solved.
During the spectral calibration, the result of the central wavelengths of different sample points shows a nonuniform result. The nonuniformity should be caused by the feature of the filter based on the F-P structure, which makes the central wavelength shift to shortwave while the incident angle grows larger. To obtain more accurate calibration parameters, the image is divided into multiple rings according to the radius of the certain pixel, and the central wavelength of each ring is solved separately. The result of five rings is shown in Table 2. The deviation caused by the filter is about 3–5 nm from the center of the image to the edge.
The radiometric calibration is performed in a dark room, with an integrating sphere as the light source and a calibrated ASD spectrometer as the standard radiation reference. The aperture of the camera and the fiber sensor head of the spectrometer are placed in the integrating sphere. The image of the camera and the light radiation intensity curve measured by the spectrometer are recorded correspondingly as the working current of the integrating sphere is adjusted. As the central wavelength differs in rings, the radiometric calibration coefficient is solved with rings as the unit. Radiometric correction is achieved with Equation (5).
L = α i j D N a v g i + β i j = α i j ( α k D N k + β k ) + β i j
where α i j and β i j are the absolute radiometric correction coefficients of ring i and wavelength j and α k and β k are the relative radiometric correction coefficients of pixel k. With the data recorded in the integrating sphere experiment, all coefficients in Equation (5) are solved with a curve fitting tool.

3.2.2. Multispectral Information Acquisition

With the reconstruction method and system calibration, the original radiation intensity curve of the whole scene can be reconstructed from the raw image captured by the BM3C. The multispectral information acquisition procedure of a chosen pixel on the reconstructed image involves the following steps:
  • By solving the included angle of the chosen object pixel and the light axis of each ommatidium channel, the nearest channel is picked out, and the image pixel of the object is solved according to the projection theory.
  • By checking the look-up table via the combined reconstruction method, the corresponding pixels in the other six images of the multispectral cluster are found.
  • With the spectral and radiometric calibration data, the central wavelength and the correction coefficient of all points are determined. The radiation intensity curve is solved using Equation (5).
With the solved radiation intensity curve, the spectral features of the captured ground objects can be read and analyzed. As BMCCEC can only solve the reflection curve of an area due to the blind area, the BM3C system can record the multispectral information of all pixels in the large object scene and acquire an absolute radiation intensity curve.

4. Airborne Imaging Experiment

To test the imaging capability of the BM3C and the reconstruction method, an airborne imaging experiment was performed. During the experiment, a customized six-rotor UAV with a 12 V power supply for the camera and a programmable onboard computer was used as the boarding platform of the camera. The BM3C was mounted below the UAV with a customized fixture, and the lens was aimed to the ground for remote sensing, as shown in Figure 14.
The onboard computer was connected to the flight control system of the UAV with the DJI onboard SDK and contained a recording program. The system was programmed to start recording with a set delay after detecting the take-off signal from the UAV and to stop recording when detecting the landing signal.
The imaging experiment is in the Xi’an Institute of Optics and Mechanics. During the recording procedure in the experiment, the exposure time of the camera was set to 3000 μs, and the acquisition framerate was set to 2 fps. According to the data of the flight control system, the camera recorded at a maximum altitude of 75 m, and the horizontal flight range was 100 m.

5. Results

In the airborne imaging experiment, 795 images were recorded during a flight procedure. The images are reconstructed with the method described in Section 3. Figure 15 shows a series of multispectral images reconstructed from one of the raw images.
The inverted radiation of four typical ground objects is sampled to evaluate the imaging result. In Figure 16, the absolute radiation intensity curves of four typical ground objects are picked from an image at a higher altitude. As shown in Figure 16b, the curve of red and green plants shows the typical response curve of the plant, which has an absorption band of around 600 nm to 650 nm and a high reflection at the waveband ranging from 750 to 800 nm, confirming the different intensities of 500–600 nm. Also, the curves of the same kind of target in a different area of the FOV show a high uniformity.
The spectrum curves of the test points in Figure 16 are reasonable and accord with the reflection characteristics of the objects. An object classification based on the k-means clustering algorithm is performed to clearly show the application potential. The result of the object classification is shown in Figure 17.
As can be seen clearly, the ground objects are classified into three categories: plants, a high-reflectivity concrete/building and a low-reflectivity concrete/wet road. Therefore, the prototype BM3C and the reconstruction method show the capacity for ground target recognition. The NDVI of the scene can be solved with the clustering result. With training data and a convolutional neural network, more detailed classification can be performed with the camera.

6. Discussion

In summary, aerial multispectral imaging with a prototype BM3C was demonstrated in this work. The camera contains 169 ommatidia distributed on a curved structure and achieves a maximum FOV of 120 degrees and a maximum seven-wavelength multispectral imaging FOV of above 98 degrees. Compared with the former BMCCEC prototype, the camera proposed in this article achieves the multispectral imaging of all pixels in the FOV with no blind area in the image (Figure 18).
A comparison of the BM3C prototype and the state-of-the-art aerial multispectral cameras is listed in Table 3 In comparison with previous research, our BM3C has a unique curved compound eye structure with multispectral imaging ability, preserving the larger FOV and more multispectral wavebands at the same time. However, the prototype system also shows a larger size and weight, which can be further reduced by changing the structure material and using a more compact design.
A comparison of the BM3C prototype and the state-of-the-art compound eye multispectral systems is listed in Table 4. With the unique multispectral lens array arrangement and the optical relay sub-system design, BM3C can achieve a large FOV by spreading out cluster units on a curved shell and a high resolution by optical optimization. Thus, the BM3C shows apparent advantages, especially in terms of the field of view, the focal length, the imaging resolution and the number of multispectral wavebands in comparison with others.
According to the results of this article, future study of the system can be undertaken in the following areas:
  • Calibration of a large number of imaging channels. In the compound eye imaging system, normal calibration methods for array cameras perform relatively poorly because of two reasons: the low imaging resolution of each ommatidium and the hardship in the parameter optimization from the large number of the imaging channels. An appropriate calibration method for the system should give rise to a better image registration result.
  • Multidimension information sensing. As the compound eye imaging system shows the capability of multispectral imaging, the system can also be used for multidimension information sensing. For example, by attaching a polarizer with different polarization angles to the ommatidia, the system can capture the polarization information of different polarization angles for navigation. This may bring on many new applications of the compound eye system.
  • Lightening and miniaturizing the camera. As shown in Table 3, the BM3C achieves a large FOV and more spectrum channels than common aerial multispectral cameras but a larger volume and weight. In the future, the system will be lightened and miniaturized via multiple ways, including using lighter materials like resin for lenses and carbon fiber for the shell and using aspheric lenses for a more compact optical system.
  • System application. Like other multispectral imaging devices, BM3C can be applied to fields like the parameters analyzing of crops or vegetation and the researching of biomass or biocommunities in a large area.

Author Contributions

Idea and methodology, W.Y. and Y.Z.; optical design and simulation, H.X. and X.Z.; structure design, fabrication and assembling, D.W., H.X. and Y.Z.; experiment, Y.Z. and Y.L.; software, Y.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, W.Y.; project administration, W.Y.; funding acquisition, W.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (61975231, 62061160488) and the China National Key Research and Development Program (2021YFC2202002).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to funders’ policy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Marris, E. Fly, and bring me data. Nature 2013, 498, 156–158. [Google Scholar] [CrossRef]
  2. Yang, G.J.; Liu, J.G.; Zhao, C.J.; Li, Z.H.; Huang, Y.B.; Yu, H.Y.; Xu, B.; Yang, X.D.; Zhu, D.M.; Zhang, X.Y.; et al. Unmanned Aerial Vehicle Remote Sensing for Field-Based Crop Phenotyping: Current Status and Perspectives. Front. Plant Sci. 2017, 8, 26. [Google Scholar] [CrossRef] [PubMed]
  3. Tanida, J.; Kumagai, T.; Yamada, K.; Miyatake, S.; Ishida, K.; Morimoto, T.; Kondou, N.; Miyazaki, D.; Ichioka, Y. Thin observation module by bound optics (TOMBO): An optoelectronic image capturing system. In Proceedings of the Conference on Optics in Computing 2000, Univ Laaval, Quebec City, QC, Canada, 18–23 June 2000; pp. 1030–1036. [Google Scholar]
  4. Zhao, Z.F.; Liu, J.; Zhang, Z.Q.; Xu, L.F. Bionic-compound-eye structure for realizing a compact integral imaging 3D display in a cell phone with enhanced performance. Optics Letters 2020, 45, 1491–1494. [Google Scholar] [CrossRef] [PubMed]
  5. Deng, H.X.; Gao, X.C.; Ma, M.C.; Li, Y.Y.; Li, H.; Zhang, J.; Zhong, X. Catadioptric planar compound eye with large field of view. Opt. Express 2018, 26, 12455–12468. [Google Scholar] [CrossRef] [PubMed]
  6. Ueno, R.; Suzuki, K.; Kobayashi, M.; Kwon, H.; Honda, H.; Funaki, H. Compound-Eye Camera Module as Small as 8.5 × 8.5 × 6.0 mm for 26 k-Resolution Depth Map and 2-Mpix 2D Imaging. IEEE Photonics J. 2013, 5, 6801212. [Google Scholar] [CrossRef]
  7. Duparré, J.; Dannberg, P.; Schreiber, P.; Bräuer, A.; Tünnermann, A. Artificial apposition compound eye fabricated by micro-optics technology. Appl. Optics 2004, 43, 4303–4310. [Google Scholar] [CrossRef]
  8. Li, L.; Yi, A.Y. Development of a 3D artificial compound eye. Opt. Express 2010, 18, 18125–18137. [Google Scholar] [CrossRef]
  9. Floreano, D.; Pericet-Camara, R.; Viollet, S.; Ruffier, F.; Bruckner, A.; Leitel, R.; Buss, W.; Menouni, M.; Expert, F.; Juston, R.; et al. Miniature curved artificial compound eyes. Proc. Natl. Acad. Sci. USA 2013, 110, 9267–9272. [Google Scholar] [CrossRef]
  10. Wu, D.; Wang, J.N.; Niu, L.G.; Zhang, X.L.; Wu, S.Z.; Chen, Q.D.; Lee, L.P.; Sun, H.B. Bioinspired Fabrication of High-Quality 3D Artificial Compound Eyes by Voxel-Modulation Femtosecond Laser Writing for Distortion-Free Wide-Field-of-View Imaging. Adv. Opt. Mater. 2014, 2, 751–758. [Google Scholar] [CrossRef]
  11. Shi, C.Y.; Wang, Y.Y.; Liu, C.Y.; Wang, T.S.; Zhang, H.X.; Liao, W.X.; Xu, Z.J.; Yu, W.X. SCECam: A spherical compound eye camera for fast location and recognition of objects at a large field of view. Opt. Express 2017, 25, 32333–32345. [Google Scholar] [CrossRef]
  12. Xu, H.R.; Zhang, Y.J.; Wu, D.S.; Zhang, G.; Wang, Z.Y.; Feng, X.P.; Hu, B.L.; Yu, W.X. Biomimetic curved compound-eye camera with a high resolution for the detection of distant moving objects. Opt. Lett. 2020, 45, 6863–6866. [Google Scholar] [CrossRef] [PubMed]
  13. Liu, J.H.; Zhang, Y.J.; Xu, H.R.; Yu, W.X. Large field of view 3D detection with a bionic curved compound-eye camera. In Proceedings of the Conference on AOPC-Optical Sensing and Imaging Technology, Beijing, China, 20–22 June 2021. [Google Scholar]
  14. Labhart, T.; Meyer, E.P. Detectors for polarized skylight in insects: A survey of ommatidial specializations in the dorsal rim area of the compound eye. Microsc. Res. Tech. 1999, 47, 368–379. [Google Scholar] [CrossRef]
  15. Awata, H.; Wakakuwa, M.; Arikawa, K. Evolution of color vision in pierid butterflies: Blue opsin duplication, ommatidial heterogeneity and eye regionalization in Colias erate. J. Comp. Physiol. A-Neuroethol. Sens. Neural Behav. Physiol. 2009, 195, 401–408. [Google Scholar] [CrossRef] [PubMed]
  16. Tanida, J.; Shogenji, R.; Kitamura, Y.; Yamada, K.; Miyamoto, M.; Miyatake, S. Color imaging with an integrated compound imaging system. Opt. Express 2003, 11, 2109–2117. [Google Scholar] [CrossRef]
  17. Miyatake, S.; Shogenji, R.; Miyamoto, M.; Nitta, K.; Tanida, J. Thin observation module by bound optics (TOMBO) with color filters. In Proceedings of the Conference on Sensors and Camera Systems for Scientific, Industrial, and Digital Photography Applications V, San Jose, CA, USA, 19–21 January 2004; pp. 7–12. [Google Scholar]
  18. Mathews, S.A. Design and fabrication of a low-cost, multispectral imaging system. Appl. Optics 2008, 47, F71–F76. [Google Scholar] [CrossRef]
  19. Kagawa, K.; Fukata, N.; Tanida, J. High-speed multispectral three-dimensional imaging with a compound-eye camera TOMBO. In Optics and Photonics for Information Processing IV; SPIE: Bellingham, DC, USA, 2010; Volume 7797. [Google Scholar] [CrossRef]
  20. Shogenji, R.; Kitamura, Y.; Yamada, K.; Miyatake, S.; Tanida, J. Multispectral imaging system by compact compound optics. In Proceedings of the Conference on Nano- and Micro-Optics for Information Systems, San Diego, CA, USA, 3–4 August 2003; pp. 93–100. [Google Scholar]
  21. Kagawa, K.; Yamada, K.; Tanaka, E.; Tanida, J. A three-dimensional multifunctional compound-eye endoscopic system with extended depth of field. Electr. Commun. Jpn. 2012, 95, 14–27. [Google Scholar] [CrossRef]
  22. Yoshimoto, K.; Yamada, K.; Sasaki, N.; Takeda, M.; Shimizu, S.; Nagakura, T.; Takahashi, H.; Ohno, Y. Evaluation of a compound eye type tactile endoscope. In Proceedings of the Conference on Endoscopic Microscopy VIII, San Francisco, CA, USA, 3–4 February 2013. [Google Scholar]
  23. Nakanishi, T.; Kagawa, K.; Masaki, Y.; Tanida, J. Development of a Mobile TOMBO System for Multi-spectral Imaging. In Proceedings of the Fourth International Conference on Photonics Solutions, Chiang Mai, Thailand, 20–22 November 2020; Volume 11331. [Google Scholar] [CrossRef]
  24. Yu, X.D.; Liu, C.Y.; Zhang, Y.J.; Xu, H.R.; Wang, Y.Y.; Yu, W.X. Multispectral curved compound eye camera. Opt. Express 2020, 28, 9216–9231. [Google Scholar] [CrossRef]
  25. Zhang, Y.J.; Xu, H.R.; Guo, Q.; Wu, D.S.; Yu, W.X. Biomimetic multispectral curved compound eye camera for real-time multispectral imaging in an ultra-large field of view. Opt. Express 2021, 29, 33346–33356. [Google Scholar] [CrossRef]
  26. Blackburn, G.A. Hyperspectral remote sensing of plant pigments. J. Exp. Bot. 2007, 58, 855–867. [Google Scholar] [CrossRef]
  27. Liu, J.H.; Zhang, Y.J.; Xu, H.R.; Wu, D.S.; Yu, W.X. Long-working-distance 3D measurement with a bionic curved compound-eye camera. Opt. Express 2022, 30, 36985–36995. [Google Scholar] [CrossRef]
  28. Hossein-Nejad, Z.; Nasri, M. Image Registration based on SIFT Features and Adaptive RANSAC Transform. In Proceedings of the IEEE International Conference on Communication and Signal Processing (ICCSP), Melmaruvathur, India, 6–8 April 2016; pp. 1087–1091. [Google Scholar]
  29. Shogenji, R.; Kitamura, Y.; Yamada, K.; Miyatake, S.; Tanida, J. Multispectral imaging using compact compound optics. Opt. Express 2004, 12, 1643–1655. [Google Scholar] [CrossRef] [PubMed]
  30. Jin, J.; Di, S.; Yao, Y.; Du, R. Design and fabrication of filtering artificial-compound-eye and its application in multispectral imaging. In Proceedings of the ISPDI 2013-Fifth International Symposium on Photoelectronic Detection and Imaging, Beijing, China, 23 August 2013. [Google Scholar]
  31. Chen, J.; Lee, H.H.; Wang, D.; Di, S.; Chen, S.-C. Hybrid imprinting process to fabricate a multi-layer compound eye for multispectral imaging. Opt. Express 2017, 25, 4180–4189. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The UAV and multispectral TOMBO system by Nakanshi et al. [23].
Figure 1. The UAV and multispectral TOMBO system by Nakanshi et al. [23].
Biomimetics 08 00556 g001
Figure 2. Arrangement methods of the compound eye system: (a) quadrilaterals, (b) circles, (c) triangles.
Figure 2. Arrangement methods of the compound eye system: (a) quadrilaterals, (b) circles, (c) triangles.
Biomimetics 08 00556 g002
Figure 3. The arrangement of the multispectral filters: (a) spectrum arrangement, (b) coordinate arrangement.
Figure 3. The arrangement of the multispectral filters: (a) spectrum arrangement, (b) coordinate arrangement.
Biomimetics 08 00556 g003
Figure 4. The FOV overlap of neighbor ommatidia.
Figure 4. The FOV overlap of neighbor ommatidia.
Biomimetics 08 00556 g004
Figure 5. Simulation result of BM3C. (a) The 3D model of BM3C’s optical system, (b) the MTF simulation result of the edge channel, (c) the distortion simulation result of the edge channel.
Figure 5. Simulation result of BM3C. (a) The 3D model of BM3C’s optical system, (b) the MTF simulation result of the edge channel, (c) the distortion simulation result of the edge channel.
Biomimetics 08 00556 g005aBiomimetics 08 00556 g005b
Figure 6. (a) The front-view multispectral array and (b) the side view of the prototype BM3C system.
Figure 6. (a) The front-view multispectral array and (b) the side view of the prototype BM3C system.
Biomimetics 08 00556 g006
Figure 7. The calibration-free image reconstruction method. (a) The coordinates map, (b) the error analysis theory.
Figure 7. The calibration-free image reconstruction method. (a) The coordinates map, (b) the error analysis theory.
Biomimetics 08 00556 g007
Figure 8. The flow chart and schematic diagram of the combined image reconstruction method. (a) Flow chart of the method, (b) the schematic diagram.
Figure 8. The flow chart and schematic diagram of the combined image reconstruction method. (a) Flow chart of the method, (b) the schematic diagram.
Biomimetics 08 00556 g008
Figure 9. The registration image of a multispectral cluster unit. (a) The original image (b) overlap map of the cluster.
Figure 9. The registration image of a multispectral cluster unit. (a) The original image (b) overlap map of the cluster.
Biomimetics 08 00556 g009
Figure 10. The comparison of two reconstruction methods. (a) The reprojection method, (b) the combined method.
Figure 10. The comparison of two reconstruction methods. (a) The reprojection method, (b) the combined method.
Biomimetics 08 00556 g010
Figure 11. Imaging parallax of a pair of ommatidia.
Figure 11. Imaging parallax of a pair of ommatidia.
Biomimetics 08 00556 g011
Figure 12. The relationship between the object distance and parallax.
Figure 12. The relationship between the object distance and parallax.
Biomimetics 08 00556 g012
Figure 13. The captured image series during the spectral calibration procedure.
Figure 13. The captured image series during the spectral calibration procedure.
Biomimetics 08 00556 g013
Figure 14. The UAV used in the experiment and the mounted camera.
Figure 14. The UAV used in the experiment and the mounted camera.
Biomimetics 08 00556 g014
Figure 15. Reconstructed image of seven spectral channels of one scene. (a) The raw image, (b) reconstructed pseudo color image with enlarged people targets, (c) reconstructed image of 506, 560, 602, 653, 704, 751 and 801 nm wavelengths.
Figure 15. Reconstructed image of seven spectral channels of one scene. (a) The raw image, (b) reconstructed pseudo color image with enlarged people targets, (c) reconstructed image of 506, 560, 602, 653, 704, 751 and 801 nm wavelengths.
Biomimetics 08 00556 g015aBiomimetics 08 00556 g015b
Figure 16. The radiation intensity curves of typical targets. (a) Reconstructed image of 560 nm with sample points, (b) curves of plant targets, (c) curves of concrete targets.
Figure 16. The radiation intensity curves of typical targets. (a) Reconstructed image of 560 nm with sample points, (b) curves of plant targets, (c) curves of concrete targets.
Biomimetics 08 00556 g016aBiomimetics 08 00556 g016b
Figure 17. The object classification result with the k-means algorithm.
Figure 17. The object classification result with the k-means algorithm.
Biomimetics 08 00556 g017
Figure 18. (a) The reconstruction image of BMCCEC, (b) the reconstruction image of BM3C.
Figure 18. (a) The reconstruction image of BMCCEC, (b) the reconstruction image of BM3C.
Biomimetics 08 00556 g018
Table 1. Key parameters of the BM3C prototype.
Table 1. Key parameters of the BM3C prototype.
ParameterDesigned ValueTested Value
Number of ommatidia169169
System focal length (mm)2.72.76
Maximum FOV (degrees)120127.4
Central wavelengths (nm)500, 560, 600, 650, 700, 750, 800506, 560, 602, 653, 704, 751, 801
Spectral resolution (nm)1011.6
Maximum framerate (fps)1313
Distortion<2%<1.78%
Table 2. The central wavelength calibration result.
Table 2. The central wavelength calibration result.
Ideal WavelengthRing 1Ring 2Ring 3Ring 4Ring 5
500507.5507.1506.5505.5504.4
560561.7561.3560.5559.3557.9
600603.6603.0602.1600.7599.1
650654.9654.3653.3651.7649.9
700705.9705.4704.3702.9701.1
750752.9752.3751.3749.7747.9
800803.5802.9801.7800.1798.1
Table 3. The parameter comparison of aerial multispectral cameras.
Table 3. The parameter comparison of aerial multispectral cameras.
Product ModelSize/mmWeight/gBandsWavelengths/
nm
Resolution/
Pixels
FOV/DegreesPower/W
Micasense RedEdge120 × 70 × 501805475, 560, 668, 717, 8401280 × 96047.2 × 35.44
Parrot Sequoia59 × 41 × 281354550, 660, 735, 7901280 × 96061.9 × 48.58
Tetracam MCA6116 × 80 × 685806490, 550, 680, 720, 800, 9001280 × 102438.3 × 31.09.8
ADC lite114 × 77 × 612003560, 660, 8402048 × 153644.5 × 34.82
BM3C194 × 194 × 23224927506, 560, 602, 653, 704, 751, 8015120 × 5120Max. 127.46
Table 4. The comparison of the optical parameters of multispectral compound eye systems.
Table 4. The comparison of the optical parameters of multispectral compound eye systems.
SystemFocal Length (mm)FOV (Degree)WavebandsNumber of Units
TOMBO (2004) [29]1.3/716
TOMBO (2010) [19]2.3516 × 16225
Planar multispectral ACE [30]//436
Multi-layer compound eye [31]//412
TOMBO (2020) [23]1.550 × 5089
BM3C2.76Max.127.47169
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Xu, H.; Liu, Y.; Zhou, X.; Wu, D.; Yu, W. Advanced Biomimetic Multispectral Curved Compound Eye Camera for Aerial Multispectral Imaging in a Large Field of View. Biomimetics 2023, 8, 556. https://doi.org/10.3390/biomimetics8070556

AMA Style

Zhang Y, Xu H, Liu Y, Zhou X, Wu D, Yu W. Advanced Biomimetic Multispectral Curved Compound Eye Camera for Aerial Multispectral Imaging in a Large Field of View. Biomimetics. 2023; 8(7):556. https://doi.org/10.3390/biomimetics8070556

Chicago/Turabian Style

Zhang, Yuanjie, Huangrong Xu, Yiming Liu, Xiaojun Zhou, Dengshan Wu, and Weixing Yu. 2023. "Advanced Biomimetic Multispectral Curved Compound Eye Camera for Aerial Multispectral Imaging in a Large Field of View" Biomimetics 8, no. 7: 556. https://doi.org/10.3390/biomimetics8070556

APA Style

Zhang, Y., Xu, H., Liu, Y., Zhou, X., Wu, D., & Yu, W. (2023). Advanced Biomimetic Multispectral Curved Compound Eye Camera for Aerial Multispectral Imaging in a Large Field of View. Biomimetics, 8(7), 556. https://doi.org/10.3390/biomimetics8070556

Article Metrics

Back to TopTop