Next Article in Journal
Technology for SMD Packaging MOX Gas Sensors
Previous Article in Journal
A Reliable and Sensitive Method Using Cyclic Voltammetry for the Detection of Airborne Fungi
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Classification of Sonar Targets in Air—A Neural Network Approach †

1
Lehrstuhl für Sensorik, Friedrich-Alexander-Universität Erlangen, 91052 Erlangen, Germany
2
Department of Ecological Science, Vrije Universiteit Amsterdam, 1081 HV Amsterdam, The Netherlands
*
Author to whom correspondence should be addressed.
Presented at the Eurosensors 2018 Conference, Graz, Austria, 9–12 September 2018.
Proceedings 2018, 2(13), 929; https://doi.org/10.3390/proceedings2130929
Published: 29 November 2018
(This article belongs to the Proceedings of EUROSENSORS 2018)

Abstract

:
Ultrasonic sonar sensors are commonly used for contactless distance measurements in application areas such as automotive and mobile robotics. They can also be used to identify and classify sound reflecting objects. In the presented work, we classify simple sonar targets of different geometric shape and size. For this purpose, we built a test stand for echo measurements that facilitates defined arbitrary translation and rotation of the targets. Artificial neural networks (ANNs) with multiple hidden layers were used as classifiers and different features were evaluated. The focus was on two features that were derived from the echoes’ cross-correlation functions with their excitation chirp signals. We could distinguish different target geometries with our features and also evaluated the ANNs’ capabilities for size discrimination of targets with the same geometric shape.

1. Introduction

Air-based ultrasonic sonar sensors are often used for obstacle avoidance and navigation purposes in application areas such as automotive, factory automation as well as mobile ground and airborne robotics [1]. In addition to identifying the distance to the closest obstacle, it is also favorable to be able to classify targets, which may then be used as landmarks for navigation and positioning. We used broadband chirp signals—contrary to most other sonar applications—as their cross-correlation functions are narrow and they are better suited to deduce object features. Besides, echolocating bats are known to use broadband signals as soon as they have to resolve objects in front of vegetation [2] and it was also shown that they are able to classify different geometrical objects independent of their size [3]. As classifiers, ANNs are an option, which were used for ultrasonic targets in air [4] and for classification of spherical targets consisting of different materials in water [5]. In our work, we focus on classification of geometrically different shaped targets and the extraction of suitable sonar features.

2. Targets

We obtained and characterized ultrasound echoes from six different targets. Those were: two flat discs (30 mm and 50 mm radii, 4 mm thickness), two cylinders (30 mm and 50 mm radii, 60 mm and 100 mm heights, respectively) and two concave hemispheres (30 mm and 50 mm radii, 2 mm wall thickness). The three different geometric shapes have got characteristic acoustic fingerprints—spectrum versus rotation angle plots (see Figure 1). Hence, they should be well-distinguishable and were thus chosen. Detailed explanations regarding the target geometries’ acoustic properties can be found in [6].

3. Measurement Setup

All measurements were performed in an anechoic chamber. The measurement setup (as shown in Figure 2) consisted of a two axis translation stage, a rotation stage on top of the translation stage on which the targets were attached, a 1/4” Bruel&Kjaer measurement microphone (Type 4939-A-011) with an amplifier (G.R.A.S. 12AK), a wideband electrostatic ultrasonic speaker (Senscomp 7000 series) with a custom-built high voltage amplifier (0 V to 400 V, 0 kHz to 200 kHz) and a National Instruments data acquisition device for analog IO (NI-USB-6356, 1.25 MSa/s, 16 bit). The microphone and the speaker were mounted closely together (20 mm center distance) at the end of the x-axis translation stage. The targets, microphone and speaker were located 1 m above the floor. The setup was optimized so that there was no detectable direct echo from itself. Echoes that resulted from multiple reflections, appeared after the target echoes in the measured waveforms.

4. Measurement Procedure

For the ANN training and test data, the targets were automatically moved along a grid and also rotated—x-direction (0.5 m to 1.8 m, 0.1 m steps); y-direction (−0.15 m to 0.15 m, 0.05 m steps); angles (−60° to 60°, 15° steps). We used downward modulated wideband chirp signals (150 kHz to 20 kHz, 1 ms duration)—referred to as wb. Additionally, we also used narrowband chirp signals (nb, 52 kHz to 48 kHz, 1 ms duration) as we were also interested in the performance that may be achieved if we used a common narrowband ultrasonic sensor, such as a piezoelectric-based transducer. Chirp signals were chosen as they make it possible to gain information regarding a large portion of the spectrum from a single echo. From the complete recorded echoes, we selected a region of interest (ROI) of 3 ms for our feature calculations. The ROI positions were determined by the sound propagation distance to the target, which is known in this case. We also obtained true negative samples—non-target echoes—for proper ANN training. For this purpose, we added additional random time to the targets’ propagation delays for the non-target ROIs, in which echoes from multiple reflections off the measurement setup appear.

5. Neural Networks and Feature Engineering

We designed our ANNs with four hidden layers in order to achieve good generalization and to avoid overfitting to the training data. The hidden layers comprised 10, 5, 5 and 3 neurons, respectively. A scaled conjugate gradient backpropagation algorithm with a cross entropy error function was used for training. For more information on ANNs, see [7]. We performed supervised learning as we knew the target positions as well as the target classes from the measurement procedure and could, thus, create labeled data sets.
Tests with various features were performed, of which two are of special interest. The first of these features is the spectrogram of an echo’s cross-correlation function with its corresponding excitation chirp signal—referred to as FSP. For each target position, the according ROI was selected from the spectrogram. The resulting 2D array that contained the values, was flattened into a 1D array which was included into the final feature vector. The second feature is the logarithmic maximum value of the previously mentioned cross-correlation function, with an additional term to compensate for spherical spreading loss, which is based on the sonar equation—referred to as FTS. This feature is proportional to a rough estimate of a reflector’s target strength. In addition, we have also combined FSP and FTS—referred to as FC.
A feature vector was chosen to consist of one of the features for five neighboring measurement grid positions. This is equivalent to a sonar sensor traveling by a target and taking five measurements. We have evaluated our ANNs by analyzing the confusion matrices and have deduced precision as well as recall values from those matrices. Precision and recall values are to be maximized towards 100% and are calculated for each target class as given:
Precision = no ( true   positives ) no ( false   positives ) + no ( true   positives )
Recall = no ( true   positives ) no ( false   negatives ) + no ( true   positives )

6. Classification Results

Results for precision and recall are summarized in Table 1. It can clearly be seen that our ANNs are able to classify the general geometric target shapes in most cases. Apart from that, good classification among different reflector sizes seems to be possible as well. Size discrimination of the hemispheres works best. Most ambiguous classifications can be observed for disc shaped targets. We assume that this happens if a disc is rotated in a way that only a weak echo is reflected back—see Figure 1 for comparison. In order to determine whether this is actually the case, further investigation is planned and may lead to optimization opportunities for good performance with all target shapes and sizes. The FTS can be used alone as well, but yields a lower performance in contrast to the spectrogram features, which show the best results at the cost of larger computational effort. Use of FC, the combination of FSP and FTS, does not seem to lead to further significantly improved performance. Also, narrowband signals were tested successfully, hence the use of piezoelectric transducers may be possible in the future.
Classification was performed with a well-defined set of targets and true negatives that can be separated from the targets well. It thus remains to be investigated how robustly the system will perform if echoes from cluttered spaces are included, which may often be the case for applications outside the lab space. In this context, also the separation of echoes from multiple targets in close distance to each other is of interest and may require the inclusion of sophisticated sound source localization techniques. Here, an additional opportunity for improvement is to engineer and test more features in combination with feature evaluation tools such as the principal component analysis, which will presumably lead to a reduction of the necessary computational effort for ANN training and employment in comparison to the use of large spectrogram feature vectors.

Author Contributions

Conceptualization, P.K.K.; Methodology, P.K.K.; Software, P.K.K.; Validation, P.K.K., R.S. and S.J.R.; Formal Analysis, P.K.K.; Investigation, P.K.K.; Resources, P.K.K. and R.S.; Data Curation, P.K.K.; Writing—Original Draft Preparation, P.K.K.; Writing—Review & Editing, P.K.K., R.S. and S.J.R.; Visualization, P.K.K.; Supervision, S.J.R.; Project Administration, S.J.R.; Funding Acquisition, S.J.R.

Funding

This research received no external funding.

Acknowledgments

We especially thank Hans-Michael Günther from the “Lehrstuhl für Sensorik“ in Erlangen. He provided practical electronics support as well as designed and built the high-voltage amplifier.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kleeman, L.; Kuc, R. Sonar Sensing. In Springer Handbook of Robotics; Siciliano, B., Khatib, O., Eds.; Springer: Berlin, Germany, 2008; pp. 491–519. ISBN 978-3-540-30301-5. [Google Scholar]
  2. Siemers, B.M.; Schnitzler, H.U. Echolocation signals reflect niche differentiation in five sympatric congeneric bat species. Nature 2004, 657–661. [Google Scholar] [CrossRef] [PubMed]
  3. Von Helversen, D. Object classification by echolocation in nectar feeding bats: Size-independent generalization of shape. J. Comp. Physiol. A 2004, 515–521. [Google Scholar] [CrossRef] [PubMed]
  4. Barshan, B.; Ayrulu, B.; Utete, S.W. Neural network-based target differentiation using sonar for robotics applications. IEEE Trans. Robot. Autom. 2000, 435–442. [Google Scholar] [CrossRef]
  5. Dmitrieva, M.; Valdenegro-Toro, M.; Brown, K.; Heald, G.; Lane, D. Object classification with convolution neural network based on the time-frequency representation of their echo. In Proceedings of the IEEE 27th International Workshop on Machine Learning for Signal Processing (MLSP), Tokyo, Japan, 25–28 September 2017. [Google Scholar] [CrossRef]
  6. Simon, R.; Holderied, M.W.; Von Helversen, O. Size discrimination of hollow hemispheres by echolocation in a nectar feeding bat. J. Exp. Biol. 2006, 3599–3609. [Google Scholar] [CrossRef] [PubMed]
  7. Bishop, C.M. Pattern Recognition and Machine Learning; Jordan, M., Kleinberg, J., Scholkopf, B., Eds.; Springer: New York, NY, USA, 2006; ISBN 978-0387-31073-2. [Google Scholar]
Figure 1. Echo fingerprint for different target geometric shapes at a distance of 1 m, relative to speaker frequency characteristic at a distance of 1 m, measured with single-frequency sine bursts.
Figure 1. Echo fingerprint for different target geometric shapes at a distance of 1 m, relative to speaker frequency characteristic at a distance of 1 m, measured with single-frequency sine bursts.
Proceedings 02 00929 g001
Figure 2. Measurement setup.
Figure 2. Measurement setup.
Proceedings 02 00929 g002
Table 1. Comparison of precision and recall for different signals and features; format: (precision in %, recall in %), wb: wideband, nb: narrowband.
Table 1. Comparison of precision and recall for different signals and features; format: (precision in %, recall in %), wb: wideband, nb: narrowband.
ShapeNo TargetDiscCylinderHemisphere
wb, FSP98.8, 10094.5, 74.184.6, 92.591.5, 100
nb, FSP97.6, 98.896.1, 90.099.2, 10098.3, 100
wb, FTS96.2, 98.695.3, 87.088.8, 97.997.0, 98.1
nb, FSP94.1, 96.781.9, 70.867.1, 97.096.3, 63.6
wb, FTS99.7, 99.797.1, 98.098.3, 97.499.1, 99.1
nb, FSP99.7, 98.695.8, 97.598.2, 95.094.0, 100
dim.r = 30mmr = 50mmr = 30mmr = 50mmr = 30mmr = 50mm
wb, FSP98.8, 10083.3, 70.295.3, 69.573.1, 92.5100, 92.698.2, 10086.3, 100
nb, FSP97.6, 98.891.3, 76.480.7, 83.698.3, 93.493.5, 10096.3, 100100, 100
wb, FTS96.2, 98.669.6, 27.154.2, 78.993.5, 95.685.2, 10093.8, 96.8100, 77.1
nb, FSP94.1, 96.743.3, 25.549.1, 57.861.2, 98.175.9, 95.785.7, 29.072.9, 72.9
wb, FTS99.7, 99.784.4, 77.682.8, 90.698.2, 90.391.4, 98.198.2, 98.2100, 100
nb, FSP99.7, 98.682.8, 84.279.0, 80.3100, 89.596.9, 100100, 10090.3, 100
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kroh, P.K.; Simon, R.; Rupitsch, S.J. Classification of Sonar Targets in Air—A Neural Network Approach. Proceedings 2018, 2, 929. https://doi.org/10.3390/proceedings2130929

AMA Style

Kroh PK, Simon R, Rupitsch SJ. Classification of Sonar Targets in Air—A Neural Network Approach. Proceedings. 2018; 2(13):929. https://doi.org/10.3390/proceedings2130929

Chicago/Turabian Style

Kroh, Patrick K., Ralph Simon, and Stefan J. Rupitsch. 2018. "Classification of Sonar Targets in Air—A Neural Network Approach" Proceedings 2, no. 13: 929. https://doi.org/10.3390/proceedings2130929

Article Metrics

Back to TopTop