Next Article in Journal
Impact of Experimental Parameters on Cell–Cell Force Spectroscopy Signature
Previous Article in Journal
On the Impact of Biceps Muscle Fatigue in Human Activity Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Do-It-Yourself Hyperspectral Imager Brought to Practice with Open-Source Python

by
Kimmo Aukusti Riihiaho
*,
Matti Aleksanteri Eskelinen
and
Ilkka Pölönen
Faculty of Information Technology, University of Jyväskylä, P.O. Box 35, FI-40014 Jyväskylä, Finland
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(4), 1072; https://doi.org/10.3390/s21041072
Submission received: 10 December 2020 / Revised: 27 January 2021 / Accepted: 29 January 2021 / Published: 4 February 2021
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Commercial hyperspectral imagers (HSIs) are expensive and thus unobtainable for large audiences or research groups with low funding. In this study, we used an existing do-it-yourself push-broom HSI design for which we provide software to correct for spectral smile aberration without using an optical laboratory. The software also corrects an aberration which we call tilt. The tilt is specific for the particular imager design used, but correcting it may be beneficial for other similar devices. The tilt and spectral smile were reduced to zero in terms of used metrics. The software artifact is available as an open-source Github repository. We also present improved casing for the imager design, and, for those readers interested in building their own HSI, we provide print-ready and modifiable versions of the 3D-models required in manufacturing the imager. To our best knowledge, solving the spectral smile correction problem without an optical laboratory has not been previously reported. This study re-solved the problem with simpler and cheaper tools than those commonly utilized. We hope that this study will promote easier access to hyperspectral imaging for all audiences regardless of their financial status and availability of an optical laboratory.

1. Introduction

Hyperspectral imaging based on unmanned aerial vehicles (UAVs), especially drones, has gain a lot of attention in recent years, for example, in precision agriculture [1,2] and precision forestry [3,4,5,6]. Compared to more traditional aircraft- and spacecraft-based operations, drones offer ease of operation, cost-efficiency, and small ground-pixel size. In addition, drone-based imaging is possible even in cloudy weather [2].
Commercial hyperspectral imagers (HSIs) are expensive, which restricts their usage for low-budget operators, such as individual researchers or landowners, as even the cheapest HSIs cost tens of thousands of euros [7,8]. Commercial HSIs are also heavy and bulky, which introduces additional requirements for drones utilized in UAV campaigns [8]. Several cheap HSIs have been introduced in the literature: some are made lightweight for UAV operations [8,9], while others target laboratory usage [7]. While several of these HSIs can be manufactured in do-it-yourself manner, calibration and aberration correction methods usually require access to an optical laboratory.
Our study builds on do-it-yourself HSI presented by Sigernes et al. [9], which we call the Sigernes design. It is a push-broom type HSI that can be manufactured with off-the-shelf optical components and cameras, which are embedded in a 3D-printed casing. Its manufacturing process does not require access to special equipment other than a thermoplastic 3D printer. A photograph of our adaptation of the Sigernes design is shown in Figure 1.
A push-broom HSI forms a hyperspectral image by scanning the area being imaged line by line. The front lens sees a circular area, which is restricted to a line by a narrow slit. Light passing through the slit is collimated before diffracting it into its wavelengths by a diffraction grating. Finally, diffracted light is focused to the sensor of a video camera. In each captured frame, each row represents a spectrograph of a slit pixel. Stacking all scanned frames together forms the final hyperspectral image, often called an image cube. The image cube then consists of two spatial dimensions and one spectral dimension.
Manual manufacturing process of the Sigernes design gives rise to a certain aberration that we call the tilt. The tilt originates from misalignment of the slit, i.e., as the slit is cut into a circular plate of metal that is inserted by hand, it is very difficult to align it perfectly with the grating and the camera. In our experience, 1–2 degree accuracy can be achieved with reasonable effort. As the tilt is specific for this design, we did not find any suggestions how to correct it in the literature, although Sigernes et al. [9] noted its existence.
Aside from the problems caused by the manufacturing process, there exist several optical aberrations that are unavoidable in push-broom imagers in general. On an ideal push-broom HSI, given an evenly illuminated area of single material, any spectral emission lines would form perfectly vertical lines of even length on the sensor. On real devices, the lines are not straight or of even length, if not corrected by optical design or software solutions [10].
Spectral smile is an aberration that causes the emission lines to have a curved shape. It is caused by dispersion when the slit image is spread over the sensor [11]. Figure 2 shows an example of a raw, uncorrected frame acquired from our imager side by side with an ideal synthetic frame that is free of aberrations. In addition, the well-known keystone aberration is barely visible in the real frame. It causes the emission lines to shorten towards the red end (right-hand side in the figure) of the spectrum. Tilt and smile aberrations are more easily distinguishable in figures in Section 3.
Optical aberrations are usually expressed in terms of point spread functions (PSFs) [10,12], but determining a PSF for a given instrument is not a trivial task and requires laboratory equipment, such as a monochromator [12] or spectral calibration lamps [13]. There exist several corrections for spectral smile [11,12,13,14], but we found none that could be carried out without an optical laboratory. For illustration of the most common optical aberrations present in push-broom imagers and suggestions on how to minimize their effect by optical design, please refer to [10]. For correction of keystone and smile of the Sigernes design when optical laboratory is available, please refer to [13].
In Section 2, we provide an improved casing for the Sigernes design as a printable 3D model (available at the project’s OSF repository at https://osf.io/3uhkb) along with an open-source software artifact capable of imaging and aberration correction (available at https://github.com/silmae/desmiler). Section 3 presents two experiments: the first one shows the effectiveness of our correction method with synthetic data, and the second one demonstrates imaging without an optical laboratory (dataset for experiment replication is also available at the OSF repository). The results of the experiments are shown in Section 4 and discussed in Section 5. Section 6 concludes the study.

2. Materials and Methods

The working principle of the Sigernes design is quite simple: light is first gathered and restricted to a narrow beam, which is then diffracted over the video camera’s sensor. The 3D-printed casing keeps the optical components immobile relative to each other and prevents stray light entering the sensor. The design of the optics is thoroughly described in [9]. Here, we highlight some changes and improvements made to the original casing. The presented 3D models are available at https://osf.io/3uhkb. Modifiable versions of the models can be opened and modified using the open-source computer aided design (CAD) software FreeCAD (https://www.freecadweb.org).
The slit needs to be aligned with the grating and the camera sensor in order to obtain correct spectral data. Any misalignment causes imaged emission lines to be tilted. In the original casing in [9], the slit could rotate freely, which we found difficult to control. Our casing fixes the slit in place when the casing is closed, which provides consistent slit orientation over imaging sessions. Furthermore, fixing the slit part of the S-tubes (2 in Figure 1), along with the collimator lens part (3 in Figure 1), allows adjustment of the focus tube during operation. A thumb hole was cut to the casing for the purpose.
Fixing holes in the bottom half of the casing fit to optical breadboards with 50 mm hole spacing for easy integration into optical laboratory equipment. For tripod usage, an adapter plate was devised (printable 3D model available at https://osf.io/srzgf).
The used video camera was Grasshopper3 model GS3-U3-91S6M-C (Flir) machine vision camera with ICX814 (Sony) monochromatic CCD sensor of 3376 × 2704 pixels. The 3D model of the casing must be modified if a different camera is being used. The camera heats up during operation, so a connection to air must be established to allow heat to dissipate. The orientation of the camera is forced to be parallel with the grating by the casing, so problems in orientating the camera with the slit image, as mentioned in [15], are avoided.
The used optical components are identical to those used in [9]. They are listed in Appendix A for completeness. From a monetary point of view, the used system costs less than 2000 euros, of which the machine vision camera costs roughly 1000 euros. Using cheaper camera and diffraction grating can bring the total price down to few hundred euros [8]. Recent development in 3D-printing imaging quality lenses [16] may lower the cost further in the future. A further benefit of using 3D-printed optics would be that, as it does not matter if normal or aberration corrected lens is printed, one could use better lenses in do-it-yourself imagers without additional cost and thus reduce the amount of aberrations.
The presented software artifact (available at https://github.com/silmae/desmiler) is capable of scanning hyperspectral image cubes, correcting spectral smile and tilt, and showing side-by-side comparisons of corrected and uncorrected data. For image cube comparison, there is a graphical user interface (GUI) with band selection and spectral angle mapping [17,18] views. Raw camera feed preview for tuning camera parameters and focus is also provided.
An effort was made to produce maintainable and modifiable program code, for example by using separate parameter script for dataset naming conventions and using an easily changeable interface for camera commands. We aimed for good coverage in documentation of the code. The software is written in Python programming language and it relies on publicly available libraries, most notably numpy [19], scipy [20], and xarray [21] for matrix calculations and data analysis, as well as open-source in-house camera controller library camazing (https://pypi.org/project/camazing).

2.1. Software Usage

The user interface (UI) of the software is implemented as an object of UI class used interactively through ipython [22] command shell. Instancing an object is simply done by running an ipython magic command %run ui.py, which will provide an UI object with name ui. Several synthetic examples (single frames and full image cubes) can also be generated and inspected using the UI object.
The flowchart in Figure 3 illustrates the operation of the software artifact. The only hardware operation is adjusting the focus with the help of the preview functionality. Operations used through the UI object are presented in the center column. There are several useful views helping the user with estimating the goodness of performed task (diamond shapes in the flowchart). Parameter tuning for scanning and aberration correction is done by writing desired values in a control file (right-hand column), which is reloaded from disk automatically when needed.
Preview is used to adjust the focus of the imager and cropping the raw sensor image to illuminated area. Scanning length and velocity can be given in arbitrary units, such as mm/s and mm, respectively. Acquisition overhead and exposure time affect how much time is allocated to acquire each frame and thus, the total frame count of the scan. The actual scanning time may differ if the imager cannot provide frames at requested rate, in which case the parameters need to be adjusted by increasing acquisition overhead or decreasing scanning velocity. Next, emission line location estimates and parameters for peak finding algorithm are given. After recording dark and white reference frames, aberration corrected reflectance image can be calculated.

2.2. Aberration Correction

Correcting the spectral smile is a well-researched subject (see, e.g., [11,12,13,14]), but we did not find any methods that could be applied without an optical laboratory. Our correction method is similar to that of Esmonde-White et al. [15] in that we search each row for pixels belonging to a certain emission line and transform the image so that curved lines become straight again. The major difference is that we do not need optical laboratory equipment for the correction. Our algorithm also takes the tilt effect into account and corrects both aberrations in a single pass. The main working principle is presented here, and an in-depth description is provided in Appendix B.
Our smile correction method consists of three distinguishable steps: locating the emission lines of a frame, constructing a shift matrix, and applying the shift. In emission line location, we use a single frame with sharp emission lines and run a peak-finding algorithm for every row. The ( x , y ) positions of the peaks are ordered in sets, each belonging to a single emission line. A circular arc is then fitted to each set of pixel positions and its radius and center point are used to calculate an appropriate shift along each row needed to straighten the emission lines. These per-pixel shifts are stored in a shift matrix, which is used to alter the original frame to form a new, corrected frame. Figure 4 illustrates the matter. Some numerical considerations on the circular arcs and using curvature as a metric for straightness are presented in Appendix C.
Let us assume a photon that should have been registered in pixel p = ( x p , y p ) was actually registered in pixel p = ( x p , y p ) . Note that we assume the shift occurring only in spectral dimension, so the spatial y-coordinate remains constant.
Essentially, we want to know the distance between detected and true column of the photon d = x p x p . Using the natural circle parameters, circle center ( a , b ) and radius r, which are given by the circle fitting, we get that the vertical distance between the circle center and pixel p is y p b . Using simple trigonometry, the distance from the circular arc to the line is
d = r 1 cos arcsin y p b r ,
which is illustrated in Figure 4. It is easy to see that, if the circle center resides in the lower half of the frame, the distance d will be greater near the top of the grid than in the bottom. This asymmetry corrects the tilt effect as a side product.
We assume that the radius r of the circular arc is much greater than y p b , so the arcus sine is defined. This is a valid assumption, as, if it were not true, the frame would be distorted useless. With that assumption, we can say that, if sgn ( a ) is positive, the emission line curves to the right and the correct value can be obtained from pixel ( x p + d , y p ) , otherwise from pixel ( x p d , y p ) .
Regardless of the position we move the emission line to, distances between emission lines are likely to change, which means that any wavelength calibrations must be done after the smile correction.

3. Experiments

To quantify the effectiveness of our method and show that the imager can be used without an optical laboratory, we conducted two experiments. In the first experiment, synthetic data were generated to show that our algorithm truly decreases the amount of spectral smile and the tilt. In the second experiment, a Macbeth color checker was scanned to show that hyperspectral image cubes can be acquired using just common office appliances.

3.1. Synthetic Frames

First, to construct an ideal, undistorted synthetic frame, a fluorescence light was imaged. The mean of several sensor rows was calculated to form a single smooth spectrogram, which acted as a base for the synthetic frame generation. This spectrum was then repeated on the synthetic sensor’s imaging area similar to real sensor. To generate non-identical, but similar spectra, each row was multiplied by a small random value drawn from a Gaussian distribution. In addition, random noise drawn from uniform distribution was added to each pixel value. The rest of the sensor rows, which would not get illuminated in the real sensor, were filled with random noise drawn from uniform distribution. The process is similar to the one used in [15]. A side-by-side comparison of a full synthetic frame and a real frame can be seen in Figure 2. For the rest of this section, we use only the illuminated (cropped) area of the sensor, as we would in the case of real frames.
The tilt and the spectral smile were generated using the correction algorithm in reverse, i.e., first generating a distortive shift matrix and interpolating each row with it. The tilt was set to 1 and the curvature to 3 × 10 5 1 px , which are common values found in real frames acquired by our imager. Examples of the generated frames cropped to the imaging area with and without distortions can be seen in Figure 5.
For statistical analysis, 1000 frames were generated. The smile correction procedure was run for each frame, and the emission line search and curve fitting was used to determine the amount of tilt and spectral smile before and after the correction. Emission line search was successful for 937 frames. It is not likely that the actual correction had failed, but rather that the emission lines were not recognized well enough. However, the matter was not investigated further.
Tilt of the emission lines was estimated with linear least square error fit in ( y , x ) coordinate system; as the lines are nearly vertical, a line fit in ( x , y ) coordinates would have caused nearly infinite slope. Curvature of the emission lines was estimated using parabolic fit as in Equation (A5). Other possible curvature estimation methods are further considered in Appendix C. This experiment can be fully replicated using the program code available at https://github.com/silmae/desmiler.

3.2. Color Checker

In the color checker experiment, a Macbeth color checker (ColorChecker Classic, X-Rite) was selected as a test target. It was taped onto the side of an office locker along with a four-fold strip of common copy paper as a white reference target. The imager was placed on an office desk against a metallic ruler fastened to the desk with a clamp and some tape. Instead of using a motorized scanning platform, the imager was slid along the ruler by hand. Distance from the front lens to the target was approximately 30 cm, while the length of the scan was 14 cm. From that distance the imager could cover the squares and some of their black boundaries in across scan direction. Imaged area consisted of the red, green, and blue color tiles (4 cm × 4 cm each). 798 frames with 110 ms exposure time were recorded. Cropping was set to 2000 pixels in spectral dimension and 800 pixels in along slit direction. The scene was illuminated by two halogen lamps with a diffuser. The lamps were set above the imager on both sides, and they were immobile relative to the target. The experiment setup is shown in Figure 6.

4. Results

Comparing the ideal frame in Figure 5a with corrected frame in Figure 5e shows that our method corrects the tilt and spectral smile to imperceptible levels. The emission lines are shifted out of the original locations, as expected. Statistical results in Table 1 show that, on average, the tilt was reduced from 0.9 to 0.005 , and spectral smile from 30 × 10 6 to 1 × 10 6 . The uncorrected values in the table show the tilt is underestimated compared to the 1 tilt used in the generation process.
One may notice that the error estimate values, the standard deviation of the mean, of the corrected frames are greater than the actual values, which implies that the correction reduces the aberrations to zero in terms of the used metrics. Numerical considerations of the curvature metric, presented in Appendix C, show that the curvature metric begins to fluctuate when values decrease below 10 6 , so this behavior is expected.
The correction was based on four emission lines covering the range approximately from 400 to 600 nm. Emission Line 3 shows significantly greater error estimates than the other three lines. We suspect that the algorithm had troubles recognizing the line properly on all occasions, but the matter was not investigated further.
The correction shifted each emission line by 5 pixels on average on the used four emission lines. The shift of lines other than those used for the correction was not investigated, but some of the emission lines in Figure 5 seem to have shifted considerably more than this. The mere existence of the location shift implies that any wavelength calibration must be performed after the smile correction and not before.
The false color reconstruction of the color checker scan in Figure 7 shows that the imager can provide decent scan quality without using optical laboratory. The black boundary between red and green tile is somewhat blurry, which is caused by the imager not being fully connected to the ruler. Vertical discontinuities near the right edge of the green tile are probably caused by the imager not being fully connected to the table during the scan. The faint horizontal stripes originate from unevenness of the slit.

5. Discussion

Our aberration correction method reduced both the spectral smile and tilt to zero within the accuracy of used metrics. The least squares fitting used as the core of both metrics sets the limit of the accuracy that can be achieved. The results should be further verified using well-established calibration methods, which is impossible with the equipment available in our laboratory.
Using the imager completely without laboratory equipment was shown to be possible, but using at least a motorized scanning platform is recommended for better quality scan results. In addition, as imaging in low lighting requires a long scanning time for reasonable signal-to-noise ratio, driving the scan manually can be quite straining. Using spectral calibration lamps with well-separated emission lines for light reference would lessen the manual labor required to find good emission line location estimates.
The software artifact is relatively easy to use, even if the user interface (UI) is a bit of a patchwork combining graphical elements, ipython console, and text file modification. However, using a very light UI makes it easy to add new features to the software and change the existing ones. The software was written in the viewpoint of correcting the spectral smile and lacks many features full imaging software should have; most notably, it lacks wavelength calibration, which is not essential for smile correction.
We made an effort to construct the software code in understandable and maintainable manner for ease of modification for different purposes. One such purpose might be to use this study as a student project for university level physics and information technology courses. For other than scientific audiences, we hope that easy-to-use software makes hyperspectral imaging more accessible and lessens the anxiety to try out hyperspectral imaging for various tasks that it can be useful in. The presented correction method removes the need for accessing an optical laboratory for spectral smile correction, and thus improves the image quality that can be expected from such low-cost imagers. We hope that this helps, in part, in adaptation of hyperspectral imaging as common tool rather than something only selected few with great resources can utilize.
For further research, the same circle fitting idea should be investigated in correcting the keystone aberration as well, as, even if it is a linear effect in itself, combined with the scaling effect that causes the red-end emission lines stretch in spatial dimension, the ends of the emission lines can be considered to form an arc. This is barely visible in Figure 2 as a sideways hourglass shape. We will continue to develop the software in the hope of eventually achieving full open-source imaging software beneficial to a greater group of users and adaptable to a large set of imagers.

6. Conclusions

In this study, we showed that spectral smile aberration of the push-broom hyperspectral imager presented by Sigernes et al. [9] can be corrected without using an optical laboratory. We also corrected an aberration, which we named tilt. The tilt is caused by the misalignment of the entrance slit and is specific for this particular imager design. We demonstrated that the imager can be used in acquisition of hyperspectral image cubes without access to an optical laboratory.
We provided simple open-source imaging software with the ability to correct spectral smile and the tilt and inspect the results. The software contains several built-in examples of spectral smile and tilt aberrations, and different kinds of examples can be generated by varying involved parameters. The program code is made publicly available through Github repository at https://github.com/silmae/desmiler under MIT license.
For those readers who wish to build their own imager, we provided required 3D models of the casing, in both print-ready and modifiable formats through https://osf.io/3uhkb. The repository also contains the dataset (hyperspectral image cube and reference frames for corrections) for reproducing the presented color checker experiment.

Author Contributions

Conceptualization, K.A.R., M.A.E., and I.P.; methodology, K.A.R., M.A.E., and I.P.; software, K.A.R.; validation, K.A.R.; writing—original draft preparation, K.A.R.; writing—review and editing, K.A.R.; visualization, K.A.R.; supervision, M.A.E. and I.P.; project administration, I.P.; and funding acquisition, I.P. All authors have read and agreed to the published version of the manuscript.

Funding

This study was partly funded by Academy of Finland (327862) and Jane and Aatos Erkko Foundation (170015).

Data Availability Statement

The program code is available at https://github.com/silmae/desmiler under MIT license. The 3D models and dataset for the color checker experimet are available at https://osf.io/3uhkb under CC BY 4.0 license.

Acknowledgments

We would like to thank all anonymous reviewers for their valuable advice and criticism. We also thank Anna-Maria Raita-Hakola for the photograph of the imager.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Appendix A. Optical Components

Table A1 lists optical components needed to build used Sigernes design push-broom hyperspectral imager.
Table A1. Shopping list for the HSI. Adapted from [9]. Part numbers refer to the part numbers used on Edmund Optics website https://www.edmundoptics.com/.
Table A1. Shopping list for the HSI. Adapted from [9]. Part numbers refer to the part numbers used on Edmund Optics website https://www.edmundoptics.com/.
DescriptionPart Number
Front lens f/4 Focal Length (FL) 16 mm83–107
M12 lock nut for μ-video lenses64–102
Precision air slit 25 μm × 3 mm38–558
Field lens FL = 10 mm63–519
3 × S-Mount thin lens mounts63–943
S-mount focus tube63–953
Collimator lens FL = 30 mm63–523
600 grooves/mm transmission grating 25 × 25 mm249–580
Detector lens f/2.5 FL = 25 mm56–776
C-mount to μ-video lens adapter53–675

Appendix B. Correction Algorithm Details

Our smile correction method is divided into three parts: locating the emission lines of a frame, constructing a shift matrix, and applying the shift. In emission line location, we use a single frame with sharp emission lines and run a peak-finding algorithm for every row. The ( x , y ) positions of the peaks are ordered in sets, each belonging to a single emission line. A circular arc is then fitted to each set of pixel positions and its radius and center point are used to calculate an appropriate shift along each row needed to straighten the emission lines. These per-pixel shifts are stored in a shift matrix, which is used to alter the original frame to form a new, corrected frame.

Appendix B.1. Locating the Emission Lines

To construct a shift matrix, we first need a frame with sharp emission lines to base the correction on. A peak finding algorithm is used for every row in order to find the locations of the emission lines in the frame. We used signal.find_peaks() from SciPy library’s Signal package. We want to find only the most prominent emission lines so that there is a good chance to find the same peak from every row. To aid the peak finding algorithm, a band-pass filter is generated from user-defined emission line location estimates.
Once the row-wise peak positions are known, if the number of found peaks per row does not match the number of position estimates given by the user, the row is discarded. The mismatch occurs if the peak finding algorithm found zero or more than one peak within some filter window. They need to match because we have no other information on which peak should be associated with which emission line.
The number of points per line is not critical for the performance of curve fitting algorithms, which are discussed in Appendix C. Tests with synthetic data showed that the acceptance rate of 95% yields curvature error of approximately 0.3%, while 5% acceptance rate yields 1% error. The difference is very similar for all used fitting algorithms. Tests with real data displayed similar behavior when normal and averaged frames were compared: averaged frames have significantly better acceptance rate due to reduced noise.
After every row has the same number of peaks, they are arranged into emission lines (represented by SpectralLine class in the implementation) so that the first point of each row belongs to Emission Line 1, the second point to Emission Line 2, and so on. Now, each emission line is presented by a list of ( x i , y i ) points, where i is the index of an accepted row.

Appendix B.2. Shift Matrix Construction

The next step is to construct the shift matrix S that stores the distances that each pixel has to be moved to be in their correct location in the spectral dimension. It has the same dimensions as the frame, and it is saved to disk to be used over several scanning sessions—until hardware changes are made. The shift matrix is the end-goal and the most important part of the smile correction. Each element s i , j S represents the distance needed to move along spectral dimension to find the correct value for each pixel p i , j in the original frame.
By selecting that the correct position of the line is at the furthest vertical distance from the circle center, we get a shift matrix where all values are either positive or negative. In the case of heavily tilted lines, the center point’s y-coordinate may not lie in frame’s area. It is then possible to have all of the emission lines to change location, as line l in Figure 4 will never touch the curve within frame’s area.
We establish a method for shift calculation in Equation (1) that assumes the smile to be consistent for every column of the frame. However, treating smile as a constant in the spectral dimension is not realistic as some lines are usually more heavily curved than others. Emission lines near the reference line would be well corrected, but lines further away possibly not. For example, if the correction were based on a heavily curved line, less curved lines elsewhere might get over-corrected, i.e., curved to the opposite direction.
If we can find more than one emission line from the frame, we can use the same method for each of them separately and interpolate over the whole area to form dissimilar columns to the shift matrix S. In other words, if we have two or more emission lines { l 1 , , l n } into which we can fit corresponding circular arcs as before, we know the row-wise shift distance d at n points, which can be interpolated to form per-pixel distance function d ( x ) for each row separately. The distance function is illustrated in Figure A1.
The selection of extrapolation outside detected emission lines is of matter to small and high end of the spectrum. If a decent number of lines can be detected, the extrapolation may well lie in the area where sensor’s quantum efficiency drops and is not too critical for the resulting frame.
Figure A1. Distance function d ( x ) along a single row constructed using known distances for emission lines l 1 , , l 4 . Linearly interpolated region are shown in blue, extrapolated regions in red. In this case, extrapolation just assumes the closest known value.
Figure A1. Distance function d ( x ) along a single row constructed using known distances for emission lines l 1 , , l 4 . Linearly interpolated region are shown in blue, extrapolated regions in red. In this case, extrapolation just assumes the closest known value.
Sensors 21 01072 g0a1

Appendix B.3. Applying Shift

Once the shift matrix S has been constructed, we can apply it to the original frame acquired from the camera. We tested two methods of applying the shift: lookup table (LUT) and row-wise interpolation (INTR). In LUT method, the value of a pixel p i , j of a frame is simply replaced with the one at pixel p i , j + s i , j . INTR method interpolates each row of the frame based on corresponding row in S.
Distances in the shift matrix are expressed as floating-point numbers, but matrix indices need to be integers for the lookup table method. Casting or rounding the floating point into an integer is bound to cause discontinuities in the smile-corrected frame. Figure A2 visualizes how rounding affects the shift matrix. The left-hand image shows the shifts in the way they are stored, as floating-point numbers. In the right-hand side image, the distances are rounded to nearest integer, which causes the final smile-corrected frame to express the same kind of discontinuities and can be seen as jaggedness of emission lines. The effect is not visible when inspecting the image cube band-wise using the CubeInspector.
Differently from previous stages which need to be performed only once per system setup, the shift must be applied for every frame. In a full spectral cube, this may mean hundreds or thousands of frames, so the effect of computational cost becomes more prominent. LUT is a relatively simple operation using Numpy library’s re-indexing capability. INTR method requires much more complicated calculations and is considerably slower. We used Xarray Python library’s xarray.dataset.interp() method for interpolation.
It is also worth noting that, when LUT method is used, some pixels at the edge of the sensor are lost, as the shifted pixel coordinates will be outside of the frame area. This does not happen with interpolation, which alters the pixel values in-place. However, interpolation usually does generate NaNs (Not a Number) that have to be cleaned away as a post-process.
Figure A2. Frame-sized shift matrices where brighter color indicates greater shift distance: (Left) floating-point presentation; and (Right) values rounded to nearest integer. Rounding causes clearly visible discontinuations.
Figure A2. Frame-sized shift matrices where brighter color indicates greater shift distance: (Left) floating-point presentation; and (Right) values rounded to nearest integer. Rounding causes clearly visible discontinuations.
Sensors 21 01072 g0a2

Appendix C. Curvature as a Metric

Smile effect causes curving in detected emission lines on the sensor, which can be clearly seen from the real frame shown in Figure 2. If we approximate a curved emission line with a circular arc, we can use reciprocal of the radius as a metric for curvature. As the emission lines are formed by a set of points, i.e., locations of spectral peaks along each sensor row, we can fit a function to that point set to gain function parameters best describing the shape.
To estimate the performance of our method, we need to be able to measure the curvature after smile correction. Fitting a circular arc to a nearly straight line accurately is not a completely trivial task [23]. We compare the accuracy and robustness of three curvature estimates produced by different fitting methods.

Appendix C.1. Least Squares Circle Fitting

One of the simplest circle fitting methods is the least squares fitting (LSF). If given n points ( x i , y i ) where i [ 1 , n ] , LSF becomes the minimization problem of the sum of squared distances from the arc. A circle can be defined by
( x a ) 2 + ( y b ) 2 = r 2 ,
where ( a , b ) is its center point and r is its radius. The unrestricted non-linear minimization problem for fitting is then
min a , b , r d i 2 ,
where d i is the Euclidean distance of each point to the circle
d i = ( x i a ) 2 + ( y i b ) 2 r .
As can be seen from this parameterization, when the arc has low curvature, i.e., large radius, the parameters become arbitrarily large, which leads to catastrophic loss of accuracy in Equation (A2) where two large and nearly equal values are subtracted from each other.

Appendix C.2. LMA Fitting

Chernov and Lesort [23] suggested changing the parameterization of the optimization problem for better stability. They suggested using Pratt’s circle parameterization [24], which defines a circle by
A x 2 + y 2 + B x + C y + D = 0 ,
where the first term describes a circle and the remaining three terms form an equation of a line. Conversion formulas to natural circle parameters a, b, and r are
A = ± 1 2 r , B = 2 A a , C = 2 A b , D = B 2 + C 2 1 4 A .
For best optimization results, Chernov and Lesort [23] suggested using unrestricted Levenberg–Maquardt corrected Gauss–Newton method with three-dimensional parameter space ( A , D , θ ) , where θ is defined by
B = 1 + 4 A D cos θ , C = 1 + 4 A D sin θ .
The distance to be minimized in Equation (A1) is
d i = 2 P i 1 + 1 + 4 A P i ,
where
P i = A x i 2 + y i 2 + 1 + 4 A D ( x i cos θ + y i sin θ ) + D .
This parameterization ensures that, even when approaching a line, none of the parameters explode. A line would just mean that A = 0 .
For partial differentials needed to form the Jacobian matrix for the optimizer, refer to the work of Chernov and Lesort [23].

Appendix C.3. Parabolic Least Squares Fitting

For a sideways opening parabola
x = a y 2 + b y + c ,
the radius of curvature at its vertex has the same length as its semi-latus rectum r = 1 2 a so the curvature is
1 r = 2 a .
This formulation does not allow arbitrary opening direction, but, assuming the tilt is only a few degrees, it does not pose a problem. Parameter b allows the vertex of the parabola to move along y-axis, thus it gives us the curvature at the point of greatest curvature. As factor a approaches zero, the equation becomes that of a line x = b y + c , which is the tangent of the parabola at the intersection with x-axis.

Appendix C.4. Optimization and Accuracy

For testing the performance of the three fitting methods and the curvature values they produce, synthetic data were generated. Generated data represent curved emission lines in x, y coordinates, where the lines are aligned along the y-axis. The x-coordinate was kept constant and small, which forces circle fitting parameters a and r to grow together when radius is increased. One hundred real frames were analyzed beforehand and extracted averages were used for synthetic data generation parameters.
A single synthetic emission line was generated by first generating 800 data points along a perfect circular arc. Their x-coordinate was then disturbed by normally distributed noise with variance σ 2 = 1.5 . Then, 95% of the points were randomly discarded using uniform distribution, which represents usual discard rate when real singular frames are acquired from the imager.
Least squares fitting was performed with scipy.optimize.leastsq() using the barycenter of the data as a starting guess for the center point.
LMA was optimized with scipy.optimize.least_squares(), which includes an implementation of Levenberg–Maquardt corrected Gauss–Newton method for unconstrained optimization problems. The result of the previously run LSF was used as a starting guess for a, b, and r, which were converted to A, D, and θ using formulas in Equation (A4).
Parabolic least squares fit was optimized with scipy.optimize.curve_fit() method, with parameters a, b, and c as in Equation (A5). A starting guess for parameter a was obtained from previously run least squares circle fit a = 1 2 r . The other starting parameters were set so that c is the mean of x-coordinates and b = 0.001 , which is essentially just an arbitrary small number.
Figure A3 shows the simulation results for emission lines generated with increasing radius. Fifty different radii were used for all three curve fitting methods. The lines represent the mean of 1000 runs and error bars as standard error of the mean are plotted for every fifth radius.
Figure A3. Synthetic emission line radius plotted against curvature acquired by different fitting methods. Parabolic fit is clearly the most stable and accurate of the three fitting methods when curvature is very low. The solid black line represents the ground truth 1 r .
Figure A3. Synthetic emission line radius plotted against curvature acquired by different fitting methods. Parabolic fit is clearly the most stable and accurate of the three fitting methods when curvature is very low. The solid black line represents the ground truth 1 r .
Sensors 21 01072 g0a3
The LSF method begins to deviate from the ground truth very early. It gives consistently too small values when r 700,000 and too great values after that. The LMA method is fairly accurate when r 500,000 and consistently overestimates the curvature thereafter. Parabolic fit is the most consistent with the ground truth. It begins to oscillate around the ground truth at approximately same radius where the LMA method starts to deviate, but it does not show any systematic deviation as the other two.
The smallest curvatures in our uncorrected experiment data were 10 5 in magnitude; thus, in terms of accuracy, the smile correction can be based on any of the three fitting methods.

References

  1. Yang, G.; Liu, J.; Zhao, C.; Li, Z.; Huang, Y.; Yu, H.; Xu, B.; Yang, X.; Zhu, D.; Zhang, X.; et al. Unmanned aerial vehicle remote sensing for field-based crop phenotyping: Current status and perspectives. Front. Plant Sci. 2017, 8, 1111. [Google Scholar] [CrossRef] [PubMed]
  2. Honkavaara, E.; Saari, H.; Kaivosoja, J.; Pölönen, I.; Hakala, T.; Litkey, P.; Mäkynen, J.; Pesonen, L. Processing and assessment of spectrometric, stereoscopic imagery collected using a lightweight UAV spectral camera for precision agriculture. Remote Sens. 2013, 5, 5006–5039. [Google Scholar] [CrossRef] [Green Version]
  3. Pölönen, I.; Annala, L.; Rahkonen, S.; Nevalainen, O.; Honkavaara, E.; Tuominen, S.; Viljanen, N.; Hakala, T. Tree Species Identification Using 3D Spectral Data and 3D Convolutional Neural Network. In Proceedings of the 2018 9th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 23–26 September 2018; pp. 1–5. [Google Scholar]
  4. Nevalainen, O.; Honkavaara, E.; Tuominen, S.; Viljanen, N.; Hakala, T.; Yu, X.; Hyyppä, J.; Saari, H.; Pölönen, I.; Imai, N.N.; et al. Individual tree detection and classification with UAV-based photogrammetric point clouds and hyperspectral imaging. Remote Sens. 2017, 9, 185. [Google Scholar] [CrossRef] [Green Version]
  5. Näsi, R.; Honkavaara, E.; Lyytikäinen-Saarenmaa, P.; Blomqvist, M.; Litkey, P.; Hakala, T.; Viljanen, N.; Kantola, T.; Tanhuanpää, T.; Holopainen, M. Using UAV-based photogrammetry and hyperspectral imaging for mapping bark beetle damage at tree-level. Remote Sens. 2015, 7, 15467–15493. [Google Scholar] [CrossRef] [Green Version]
  6. Hytönen, J.; Wall, A. Foliar colour as indicator of nutrient status of Scots pine (Pinus sylvestris L.) on peatlands. For. Ecol. Manag. 2006, 237, 156–163. [Google Scholar] [CrossRef]
  7. Stuart, M.B.; Stanger, L.R.; Hobbs, M.J.; Pering, T.D.; Thio, D.; McGonigle, A.J.; Willmott, J.R. Low-Cost Hyperspectral Imaging System: Design and Testing for Laboratory-Based Environmental Applications. Sensors 2020, 20, 3293. [Google Scholar] [CrossRef] [PubMed]
  8. Salazar-Vazquez, J.; Mendez-Vazquez, A. A plug-and-play Hyperspectral Imaging Sensor using low-cost equipment. HardwareX 2020, 7, e00087. [Google Scholar] [CrossRef]
  9. Sigernes, F.; Syrjäsuo, M.; Storvold, R.; Fortuna, J.; Grøtte, M.E.; Johansen, T.A. Do it yourself hyperspectral imager for handheld to airborne operations. Opt. Express 2018, 26, 6021–6035. [Google Scholar] [CrossRef] [PubMed]
  10. Mouroulis, P.; Green, R.O.; Chrien, T.G. Design of pushbroom imaging spectrometers for optimum recovery of spectroscopic and spatial information. Appl. Opt. 2000, 39, 2210–2220. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Jupp, D.L.B.; Datt, B.; McVicar, T.R.; Niel, T.G.V.; Pearlman, J.S.; Lovell, J.L.; King, E.A. Improving the analysis of Hyperion red-edge index from an agricultural area. In Image Processing and Pattern Recognition in Remote Sensing; Ungar, S.G., Mao, S., Yasuoka, Y., Eds.; International Society for Optics and Photonics; SPIE: Hangzhou, China, 2003; Volume 4898, pp. 78–92. [Google Scholar] [CrossRef]
  12. Skauli, T. An upper-bound metric for characterizing spectral and spatial coregistration errors in spectral imaging. Opt. Express 2012, 20, 918–933. [Google Scholar] [CrossRef] [PubMed]
  13. Henriksen, M.B.; Garrett, J.L.; Prentice, E.F.; Stahl, A.; Johansen, T.A.; Sigernes, F. Real-Time Corrections for a Low-Cost Hyperspectral Instrument. In Proceedings of the 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–5. [Google Scholar] [CrossRef] [Green Version]
  14. Dadon, A.; Ben-Dor, E.; Karnieli, A. Use of derivative calculations and minimum noise fraction transform for detecting and correcting the spectral curvature effect (smile) in Hyperion images. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2603–2612. [Google Scholar] [CrossRef]
  15. Esmonde-White, F.W.L.; Esmonde-White, K.A.; Morris, M.D. Minor Distortions with Major Consequences: Correcting Distortions in Imaging Spectrographs. Appl. Spectrosc. 2011, 65, 85–98. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Assefa, B.G.; Pekkarinen, M.; Partanen, H.; Biskop, J.; Turunen, J.; Saarinen, J. Imaging-quality 3D-printed centimeter-scale lens. Opt. Express 2019, 27, 12630–12637. [Google Scholar] [CrossRef] [PubMed]
  17. De Carvalho, O.A.; Meneses, P.R. Spectral correlation mapper (SCM): An improvement on the spectral angle mapper (SAM). In Summaries of the 9th JPL Airborne Earth Science Workshop, JPL Publication 00-18; JPL Publication: Pasadena, CA, USA, 2000; Volume 9. [Google Scholar]
  18. Yuhas, R.H.; Goetz, A.F.H.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the spectral angle mapper (SAM) algorithm. In Summaries of the Third Annual JPL Airborne Geoscience Workshop; JPL Publication: Pasadena, CA, USA, 1992; Volume 1, pp. 147–149. [Google Scholar]
  19. Harris, C.R.; Millman, K.J.; van der Walt, S.J.; Gommers, R.; Virtanen, P.; Cournapeau, D.; Wieser, E.; Taylor, J.; Berg, S.; Smith, N.J.; et al. Array programming with NumPy. Nature 2020, 585, 357–362. [Google Scholar] [CrossRef] [PubMed]
  20. Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Hoyer, S.; Hamman, J. xarray: N-D labeled arrays and datasets in Python. J. Open Res. Softw. 2017, 5. [Google Scholar] [CrossRef] [Green Version]
  22. Pérez, F.; Granger, B.E. IPython: A System for Interactive Scientific Computing. Comput. Sci. Eng. 2007, 9, 21–29. [Google Scholar] [CrossRef]
  23. Chernov, N.I.; Lesort, C. Least squares fitting of circles and lines. arXiv 2003, arXiv:cs/0301001. [Google Scholar]
  24. Pratt, V. Direct least-squares fitting of algebraic surfaces. ACM SIGGRAPH Comput. Graph. 1987, 21, 145–152. [Google Scholar] [CrossRef]
Figure 1. Our version of the Sigernes design push-broom imager with the top half removed. Light enters through the front lens (1). Connected to it is the series of S-mounts that contains the slit and the field lens (2), and the S-mount focus tube and the collimator lens (3). Diffraction grating (4) splits the light into its wavelengths. The detector lens (5) is connected to the machine vision camera (6).
Figure 1. Our version of the Sigernes design push-broom imager with the top half removed. Light enters through the front lens (1). Connected to it is the series of S-mounts that contains the slit and the field lens (2), and the S-mount focus tube and the collimator lens (3). Diffraction grating (4) splits the light into its wavelengths. The detector lens (5) is connected to the machine vision camera (6).
Sensors 21 01072 g001
Figure 2. Real frame with aberrations and ideal synthetic frame compared side by side. The used imaging target was a fluorescence lamp.
Figure 2. Real frame with aberrations and ideal synthetic frame compared side by side. The used imaging target was a fluorescence lamp.
Sensors 21 01072 g002
Figure 3. Flowchart of proposed imaging and aberration correction process. Hardware operations are located in the left-hand column, the center column holds software actions and the rightmost column represents settings written to the control file.
Figure 3. Flowchart of proposed imaging and aberration correction process. Hardware operations are located in the left-hand column, the center column holds software actions and the rightmost column represents settings written to the control file.
Sensors 21 01072 g003
Figure 4. Emission line l is shown in red, and the fitted circular arc corresponding to it is shown in blue. The light gray grid represents the sensor of the camera. In reality, the radius of the circle would be several times larger than the width of the cropped frame acquired from the sensor.
Figure 4. Emission line l is shown in red, and the fitted circular arc corresponding to it is shown in blue. The light gray grid represents the sensor of the camera. In reality, the radius of the circle would be several times larger than the width of the cropped frame acquired from the sensor.
Sensors 21 01072 g004
Figure 5. Examples of synthetic frames with distortions. As the effects are difficult to distinguish with the naked eye, the red line was added for aid. Frame (a) is an ideal frame that one would hope to get from a push-broom HSI. Frames (b,c) show the tilt and spectral smile aberrations, respectively. Frame (d) shows both aberrations combined, which represents the real data we get from the imager. Figure (e) shows Frame (d) after running the correction algorithm. The distortions are no longer distinguishable, but the emission line locations have shifted (see fourth line from the left).
Figure 5. Examples of synthetic frames with distortions. As the effects are difficult to distinguish with the naked eye, the red line was added for aid. Frame (a) is an ideal frame that one would hope to get from a push-broom HSI. Frames (b,c) show the tilt and spectral smile aberrations, respectively. Frame (d) shows both aberrations combined, which represents the real data we get from the imager. Figure (e) shows Frame (d) after running the correction algorithm. The distortions are no longer distinguishable, but the emission line locations have shifted (see fourth line from the left).
Sensors 21 01072 g005
Figure 6. The used scanning setup: Two halogen lamps with a diffusor, color checker as an imaging target, and the hyperspectral imager, which was slid along the metallic ruler fixed to the desk.
Figure 6. The used scanning setup: Two halogen lamps with a diffusor, color checker as an imaging target, and the hyperspectral imager, which was slid along the metallic ruler fixed to the desk.
Sensors 21 01072 g006
Figure 7. False color image constructed from the scan of the color checker. Scanning direction was from right to left. Blurry edges between the color tile boundaries are caused by uneven scanning speed of the manually driven scan. Vertical shift in the green tile is due to losing connection to the desk during the scan.
Figure 7. False color image constructed from the scan of the color checker. Scanning direction was from right to left. Blurry edges between the color tile boundaries are caused by uneven scanning speed of the manually driven scan. Vertical shift in the green tile is due to losing connection to the desk during the scan.
Sensors 21 01072 g007
Table 1. Mean values of tilt and curvature of 937 synthetic frames. Frames were generated with 1 tilt and 3 × 10 5   1 px curvature, which correlate well with mean values shown in original columns. Used error estimate is the standard deviation of the mean. Corrected values show that they are zero at the limit of the used metrics, as the error is greater than the values themselves. The four emission lines selected for the correction cover wavelengths approximately from 400 to 600 nm.
Table 1. Mean values of tilt and curvature of 937 synthetic frames. Frames were generated with 1 tilt and 3 × 10 5   1 px curvature, which correlate well with mean values shown in original columns. Used error estimate is the standard deviation of the mean. Corrected values show that they are zero at the limit of the used metrics, as the error is greater than the values themselves. The four emission lines selected for the correction cover wavelengths approximately from 400 to 600 nm.
ELBandTiltCurvature
OriginalCorrectedOriginalCorrected
1629 ( 902 ± 13 ) × 10 3 ( 1 ± 7 ) × 10 3 ( 29.4 ± 2.5 ) × 10 6 ( 0.1 ± 1.2 ) × 10 6
2762 ( 899 ± 19 ) × 10 3 ( 3 ± 79 ) × 10 3 ( 29.0 ± 3.5 ) × 10 6 ( 0.9 ± 10.3 ) × 10 6
3980 ( 887 ± 35 ) × 10 3 ( 13 ± 107 ) × 10 3 ( 27.3 ± 6.7 ) × 10 6 ( 3.5 ± 23.3 ) × 10 6
41517 ( 902 ± 7 ) × 10 3 ( 1 ± 5 ) × 10 3 ( 29.7 ± 1.2 ) × 10 6 ( 0.1 ± 0.8 ) × 10 6
Total mean ( 897 ± 14 ) × 10 3 ( 5 ± 14 ) × 10 3 ( 28.9 ± 2.5 ) × 10 6 ( 1.2 ± 2.6 ) × 10 6
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Riihiaho, K.A.; Eskelinen, M.A.; Pölönen, I. A Do-It-Yourself Hyperspectral Imager Brought to Practice with Open-Source Python. Sensors 2021, 21, 1072. https://doi.org/10.3390/s21041072

AMA Style

Riihiaho KA, Eskelinen MA, Pölönen I. A Do-It-Yourself Hyperspectral Imager Brought to Practice with Open-Source Python. Sensors. 2021; 21(4):1072. https://doi.org/10.3390/s21041072

Chicago/Turabian Style

Riihiaho, Kimmo Aukusti, Matti Aleksanteri Eskelinen, and Ilkka Pölönen. 2021. "A Do-It-Yourself Hyperspectral Imager Brought to Practice with Open-Source Python" Sensors 21, no. 4: 1072. https://doi.org/10.3390/s21041072

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop