Next Article in Journal
SafeSoCPS: A Composite Safety Analysis Approach for System of Cyber-Physical Systems
Previous Article in Journal
Backscatter-Assisted Collision-Resilient LoRa Transmission
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Noise Reduction Combining a General Microphone and a Throat Microphone

by
Junki Kawaguchi
and
Mitsuharu Matsumoto
*
Graduate School of Informatics and Engineering, University of Electro-Communications, 1-5-1, Chofugaoka, Chofu-shi 182-8585, Japan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(12), 4473; https://doi.org/10.3390/s22124473
Submission received: 12 May 2022 / Revised: 2 June 2022 / Accepted: 11 June 2022 / Published: 13 June 2022
(This article belongs to the Section Sensor Networks)

Abstract

:
In this study, we propose a method to reduce noise from speech obtained from a general microphone using the information of a throat microphone. A throat microphone records a sound by detecting the vibration of the skin surface near the throat directly. Therefore, throat microphones are less prone to noise than ordinary microphones. However, as the acoustic characteristics of the throat microphone differ from those of ordinary microphones, its sound quality degrades. To solve this problem, this study aims to improve the speech quality while suppressing the noise of a general microphone by using the information recorded by a throat microphone as reference information to extract the speech signal in general microphones. In this paper, the framework of the proposed method is formulated, and several experiments are conducted to evaluate the noise suppression and speech quality improvement effects of the proposed method.

1. Introduction

In recent years, with the development of information and communication devices, there has been an increase in the use of devices that use voice input, such as mobile devices, e.g., smartphones, for making calls. Although the use of voice input devices is comfortable in a noise-free environment, they may be used in a noisy environment. In such a case, the voice may be difficult to hear or unbearable to use. Therefore, speech enhancement is one of the important issues in acoustic processing to emphasize the target speech from the mixed signal.
The approach for speech enhancement can be categorized into two types. One is the approach using a single-channel signal. The other is the approach using multi-channel signals. The survey on speech enhancement classifies algorithms for speech enhancement of single-channel signals into three categories [1].
The spectral subtraction (SS) algorithm is one of those approaches. Early approaches include spectral subtraction in the correlation domain proposed by Weiss et al. [2] and that in the Fourier transform domain proposed by Boll [3]. Spectral subtraction is a powerful approach to single-channel speech enhancement and has undergone various improvements [4]. Another approach is a statistical model-based algorithm. For example, McAulay et al. used the maximum likelihood method to estimate the Fourier transform spectrum of the target signal [5]. The other approach is a subspace algorithm. Dendrinos et al. proposed an algorithm that uses singular value decomposition [6]. Ephraim et al. proposed an algorithm that uses an eigenvalue decomposition of signals [7]. The single-channel signal approach is attractive in that it uses only a single recording source. However, information that can be used for speech enhancement is limited to that related to the sound source itself. In recent years, deep learning-based approaches have also been actively studied but require a lot of computational power [8,9,10].
The microphone array is a typical method using multi-channel signals [11,12]. A microphone array uses multiple microphones to achieve noise reduction using the phase and amplitude differences of the sound entering each microphone. The microphone array is implemented in various ways. However, in principle, its implementation requires special equipment, such as a large number of microphones and complex processing. There have been various approaches to the study of noise reduction techniques. In general, noise reduction techniques using microphones have been widely studied. Well-known examples are sound focusing by phase homomorphism and adaptive microphone array by generating the zero point in the noise direction. In the field of blind source separation, independent component analysis [13,14,15] and sparsity-based source separation [16,17] are well known, but they require special equipment, such as a large number of microphones, for implementation.
One way to use voice input devices comfortably without being affected by noise is to use a microphone that is less susceptible to noise. A throat microphone is an example and records sound by detecting the vibration of the skin near the throat directly. Compared to general microphones that use air vibration to record, the throat microphone is airless and, therefore, more resistant to noise. Some studies paid attention to the characteristics and aimed to improve the accuracy of speech recognition by using a throat microphone that records by directly detecting the vibration of the skin near the throat [18]. Although the feature of throat microphones is useful, as described above, their acoustic characteristics differ from those of ordinary microphones, resulting in degraded speech quality.
In image processing fields, there are some studies on sensor fusion to obtain better image quality. Cross bilateral filter (Joint bilateral filter) is a filter technique that combines non-flash images with flash images taken at the same location [19,20]. The filter inputs are flash and non-flash images, resulting in a more natural quality image as the output. Other research cases have been reported that combine infrared and visible camera images to detect various objects, such as faces [21], pedestrians [22] and cars [23].
Inspired by these studies, we apply these concepts to the acoustical field and aim to obtain more natural speech by eliminating noise in speech signals obtained by a general microphone by using speech signals obtained by a throat microphone as a reference.
In this study, we propose a noise reduction method for general microphones based on speech signals from throat microphones. We also evaluate the effectiveness of the proposed method by conducting experiments. The rest of the paper is summarized as follows: In Section 2, we describe the typical binary mask using two microphones to clarify the proposed method. In Section 3, we describe the proposed approach and the usage scenario. In Section 4, some experiments are conducted using the proposed method to evaluate the noise reduction performance of the proposed method and the sound quality compared to a throat microphone. In Section 5, we present a discussion and conclusions on the results of the experiment and future perspectives.

2. Binary Mask Using Two Microphones

2.1. Problem Formulation

To clarify our approach, we first show the typical method using the binary mask with two general microphones [24,25,26]. A typical approach using binary masks assumes two microphones. For the mixed signal x1(t) obtained from microphone 1, removing the effects of attenuation and delay does not lose the generality of the problem. Hence, x1(t) can be written as follows:
x 1 ( t ) = s ( t ) + i = 1 N n i ( t ) ,
where s(t) is the objective signal, and ni(t) is the ith noise (i = 1, 2, 3, …, N). On the other hand, it is necessary to consider the delay and attenuation of microphone 2’s signal with respect to the signal acquired by microphone 1. Therefore, the second mixed signal x2(t) can be written as:
x 2 ( t ) = a s ( t δ ) + i = 1 N a i n i ( t δ i ) ,
where δ and δi are the delays from the objective signal and the delay from the noise j, respectively. a and aj are parameters that indicate the relative attenuation between the sensors for the objective signal and noise, respectively.
When we define Δ as the maximum possible delay between sensors, the following constraints are obtained:
| δ | Δ ,
| δ i | Δ ,
The sparse approach assumes that the target signal and noise are sparse in the time-frequency domain. Let S(τ, ω) and Ni(τ, ω) be the short-time Fourier transform of s(t) and nj(t), respectively. S(τ, ω) can be described as follows:
S ( τ , ω ) = τ = s ( t + τ ) W ( τ ) exp ( i ω τ ) ,
where τ and ω are the time frame and the angular frequency, respectively. W(τ) is the window function. If the sparsity between the objective signal and all the noise is satisfied, we can obtain the following condition:
S ( τ , ω ) N i ( τ , ω ) = 0   τ , ω ,

2.2. Noise Reduction Using Binary Mask

In order to perform denoising, the delay and attenuation parameters in Equation (2) are estimated. If Xi(τ, ω) is a short-time Fourier transform of xi(t), it can be written as follows.
X i ( τ , ω ) = t = x i ( t + τ ) W ( t ) exp ( i ω t ) ,
The signals obtained by microphones 1 and 2 can be rewritten in the time-frequency domain as follows.
[ X 1 ( τ , ω ) X 2 ( τ , ω ) ] = [ 1 1 1 a exp ( i ω δ ) a 1 exp ( i ω δ 1 ) a N exp ( i ω δ N ) ] [ S ( τ , ω ) N 1 ( τ , ω ) N N ( τ , ω ) ]
If the target signal and noise are sparse in the time-frequency domain, the following equation is obtained:
[ X 1 ( τ , ω ) X 2 ( τ , ω ) ] = [ 1 a ( τ s , ω s ) exp ( i ω s δ ( τ s , ω s ) ) ] s ( τ s , ω s )
where the objective signal exists at (τs, ωs). To estimate (τs, ωs), we calculate the ratio of X1(τs, ωs) and X2(τs, ωs). Let us define a(τs, ωs) and δ(τs, ωs) as the relative amplitude and delay parameters. We can estimate a(τs, ωs) and δ(τs, ωs) as follows:
( a ( τ s , ω s ) , δ ( τ s , ω s ) ) = ( | X 2 ( τ , ω ) X 1 ( τ , ω ) | , 1 ω X 2 ( τ , ω ) X 1 ( τ , ω ) )
where
a exp ( i ϕ ) = ϕ ,   π < ϕ < π ,
We can design the time-frequency mask M(τ, ω) for (τ, ω) as follows:
M ( τ , ω ) = { 1 , | ln a ( τ , ω ) ln a | < Δ a 2   | ln δ ( τ , ω ) ln δ | < Δ δ 2 0 ,   o t h e r w i s e
where Δa and Δδ are defined as the amplitude and delay resolution width of the histograms, respectively. In order to remove noise from the signal obtained by microphone 1, the masking process is executed as follows.
S ( τ , ω ) = M ( τ , ω ) X 1 ( τ , ω ) ,
Although the noise reduction performance of the binary mask is high, its quality depends on the accuracy of the estimated masking information. In the presence of environmental noise, the noise does not fully meet the sparsity assumptions, making it difficult to estimate mask information.

3. Proposed Approach

3.1. Overview of the Proposed Method

Figure 1 shows an overview of the proposed method. In our supposed scenario, we record the sound with a general microphone and a throat microphone simultaneously. We then create a binary mask based on the speech data obtained from the throat microphone to remove noise from the speech data obtained from the general microphone.
To achieve this goal, speech data is first captured by both a throat microphone and a general microphone. We then transformed the sound into frequency domain data by Fourier transform. The next step is to set a threshold value for the power spectrum of the speech data from the throat microphone and create a binary mask with 1 for the power above the threshold value and 0 for the rest. The audio data from the general microphone is also subjected to Fourier transform and converted to frequency domain data. The binary mask is applied to the obtained frequency data from the general microphone to remove noise. Finally, the noise-removed signal is inverse Fourier transformed to obtain the output signal.

3.2. Formulation by Equation

We formulate the proposed approach in this section. Let x1(t) be the mixed signal obtained from a general microphone and x2(t) be the signal obtained from a throat microphone. T is time. The spectra of x1(t) and x2(t) are denoted as X1(τ, ω) and X2(τ, ω), respectively, when the signals are mapped into frequency space by Fourier transform. τ is the time frame, and ω is the angular frequency. When the spectrum of the objective signal is S(τ, ω) and the spectrum of the ith noise is Ni(τ, ω), X1(τ, ω) is expressed as follows.
X 1 ( τ , ω ) = S ( τ , ω ) + i = 1 N N i ( τ , ω ) ,
On the other hand, the signal of the throat microphone does not contain noise signals, while the frequency response of the sound recorded by the throat microphone is different from that of the microphone. X2(τ, ω) is expressed as follows.
X 2 ( τ , ω ) = B ( ω ) S ( τ , ω ) ,
where B(ω) is the frequency response of a throat microphone to a general microphone. For noise reduction, a threshold is set, and a binary mask M(τ, ω) is generated as follows. th is a threshold.
M ( τ , ω ) = { 1                   | X 2 ( τ , ω ) | t h   0                   | X 2 ( τ , ω ) | < t h  
Based on the mask information, the following process is applied to the spectral signal X1(τ, ω) of a general microphone to remove noise and obtain the output signal spectrum Y(τ, ω).
Y ( τ , ω ) = M ( τ , ω ) X 1 ( τ , ω )
Finally, the obtained signal Y(τ, ω) is inverse Fourier transformed to obtain the output signal y(t). Binary mask estimation using the microphone ratio is difficult if sparse assumptions are not met.
On the other hand, our method obtains mask information directly from the throat microphone. Therefore, the proposed method can realize efficient noise removal even for a voice that does not satisfy the assumption of sparsity.
We also note that the delay between the throat microphone and the general microphone is not taken into account. The main reason is due to the difference between the typical binary mask usage scenario and the proposed method usage scenario. In the typical binary mask usage scenario, the speaker is apart from the listener, and the listener records the sound from the speaker. The proposed method aims to record the speaker’s voice clearly in a situation where the surroundings are extremely noisy. Since this work is performed on the speaker side, it is assumed that recording with a microphone and a throat microphone will be performed at the same time.

4. Experiment

4.1. Experimental Method

The effectiveness of the method proposed in Section 2 is evaluated by experiments. In this study, the accuracy of the noise removal is verified by creating the target speech, noise speech and mixed speech in advance. For the noise, we used sound effects from the Sound Effects Lab [27] and Hashimoto Research Institute [28]. Recorded speech was used as the voice. Recording with the general microphone and throat microphone was performed at the same time in a normal quiet room. The used general microphone was M4U made by inMusic, Inc. (Cumberland, RI, USA). The used throat microphone was DN-915129 made by ThirdWave Co., Ltd. (Tokyo, Japan). To evaluate the noise reduction performance quantitatively, the mixtures were created on a computer. All programs were written in Python.
The experimental conditions are shown in Table 1. Three types of noise signals were synthesized for each target signal, for a total of three types of signals. The noise level was described by dBFS. The threshold was changed in 10 dB steps for noise removal. A Hamming window was used as the window function. In this study, Signal to Distortion Ratio (SDR) [29], Signal to Noise Ratio (SNR) and Signal to Interference Ratio (SIR) [24] were used for evaluation between the target speech and the denoised speech.
The Signal to Distortion Ratio (SDR) is a measure of how much the speech is distorted relative to the target speech after noise reduction. Ŝ(τ, ω) is the signal to be evaluated for the target signal, and S(τ, ω) is the target signal. SDR is defined as follows.
S D R = 10 log 10 ( τ , ω | S ( τ , ω ) | τ , ω | | S ( τ , ω ) | λ | S ^ ( τ , ω ) | |   )
λ is a parameter used to normalize the power of the signal and is expressed as follows.
λ = τ , ω | S ( τ , ω ) | τ , ω | S ^ ( τ , ω ) |
The signal-to-noise ratio (SNR) is the ratio of the power of the signal to that of the noise, which is used to evaluate how much noise is included in the signal mixture. The higher the value of SNR is, the smaller the effect of noise is. If S(τ, ω) is the original target signal and N(τ, ω) is the noise signal, SNR is defined as follows.
S N R = 10 log 10 ( τ , ω | S ( τ , ω ) | τ , ω | N ( τ , ω ) |   )
On the other hand, the signal to interference ratio (SIR) can be used to evaluate how much noise is contained in the output after applying the binary mask [11]. Assuming that the binary mask is M(τ, ω), SIR is defined as follows:
S I R = 10 log 10 ( τ , ω | M ( τ , ω ) S ( τ , ω ) | τ , ω | M ( τ , ω ) N ( τ , ω ) |   )
By comparing these results, it is possible to understand the effect of noise on the target signal of the speech after denoising.

4.2. Quantitative Evaluation of Experiments

The SDR of the throat microphone to the signal obtained by a standard microphone was 0.474 dB in noise-free conditions. Table 2, Table 3 and Table 4 show the SDR values between the target signal and the speech after noise reduction and the SIR between the noise signal and the speech after noise reduction when three types of noise are mixed, respectively.
The experimental results are expressed in four significant digits. The background color is yellow when the SDR value is the highest. Table 2, Table 3 and Table 4 show that the SDR values of the noise-eliminated speech are larger than those of the throat microphone in all cases, indicating that the speech recorded by the noise-eliminated speech is closer to that recorded by a conventional microphone. Next, the SNRs of each noise signal and the target signal are shown in Table 5.
Noise is unlikely to be mixed into the voice recorded by the throat microphone. However, the frequency response of the sound recorded by the throat microphone is different from that of the microphone. SDR shows the similarity to clear voice recorded with a general microphone. Hence, the fact that the SDR at the output of the proposed method is higher than that of the throat microphone means that the proposed method can be used to obtain a sound closer to the sound recorded by a normal microphone than by recording with a throat microphone.
Comparing the SNR values in Table 5 with the SIR values in Table 2, Table 3 and Table 4, it was confirmed that under the same noise conditions, the speech after denoising was larger, and the effect of noise was reduced. The waveform and spectrogram of the results obtained in the experiment are shown in Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7. The spectrograms and waveforms for the maximum SDR among all the threshold values are shown in the figures. Figure 3, Figure 5 and Figure 7 also show the figure of the binary mask. It is noted that the output spectrogram is colored even in the high-frequency range, but the actual power is almost zero. This is due to the rounding error caused by outputting the spectrogram again after obtaining the output waveform once by the proposed method.
Figure 2 and Figure 3 show the waveform and the spectrogram, respectively, when intersection noise is used as noise.
Figure 4 and Figure 5 show the waveform when we used white noise whose noise level was 0 dB. Figure 6 and Figure 7 show the waveform when we used white noise whose noise level was −15 dB. As shown in Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7, even if the signal is buried in noise, the objective signal can be extracted well.

5. Discussion and Conclusions

In this study, we proposed a noise reduction method using a combination of a throat microphone and a conventional microphone. In the proposed method, a throat microphone is used as a reference signal to remove the noise in the signal obtained from a general microphone.
In the experiments, white noise and general noise were used as mixtures, and SDR and SIR were used as evaluation criteria to evaluate the noise suppression performance and speech quality. The results showed that the proposed method was able to remove noise. The proposed method applies a mask created based on the information of the throat microphone to the sound recorded by a general microphone, and it is considered that noise is mixed in the remaining frequency range. However, since voice information is sparse in the frequency domain, it is considered that the SN ratio after masking is improved even if the SN ratio is bad in the time domain. As a result, it is considered that masked sound from the general microphone has better audio quality than the sound from the throat microphone.
In the experiment, the voice of the speaker was recorded simultaneously using both a microphone and a throat microphone. Strictly speaking, there may be some time lag, but good results have been obtained experimentally without considering it.
However, there are some problems to be solved in the proposed method. For example, the frequency range in which a throat microphone can record is lower than the frequency range in which a general microphone can record. One solution is to apply the proposed method only to the frequency domain where the throat microphone can record. However, even with this approach, noise remains in the high-frequency components. The inability to record high-frequency components is a problem caused by the use of piezoelectric elements in the throat microphone. Hence, it is necessary to improve the sensor in order to substantially improve this problem.
In order to put this method to practical use, we are thinking of implementing the proposed method in real-time. We would like to proceed with experiments using the signals recorded in real noisy environments in the future. We also would like to investigate a method to estimate the threshold value automatically.

Author Contributions

Conceptualization, J.K. and M.M.; methodology, M.M.; validation, J.K.; investigation, J.K. and M.M.; writing—original draft preparation, J.K.; writing—review and editing, M.M.; supervision, M.M.; project administration, M.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Loizou, P.C. Speech Enhancement: Theory and Practice, 2nd ed.; CRC Press: London, UK, 2007. [Google Scholar]
  2. Weiss, M.; Aschkenasy, E.; Parsons, T. Study and Development of the INTEL Technique for Improving Speech Intelligibility; Technical Report NSC-FR/4023; Nicolet Scientific Corporation: Northvale, NJ, USA, 1975. [Google Scholar]
  3. Boll, S.F. Suppression of acoustic noise in speech using spectral subtraction. IEEE Trans. Acoust. Speech Signal Process 1979, 27, 113–120. [Google Scholar] [CrossRef] [Green Version]
  4. Yamashita, K.; Ogata, S.; Shimamura, T. Improved spectral subtraction utilizing iterative processing. IEICE Trans Fundametals 2005, J88-A, 1246–1257. [Google Scholar] [CrossRef]
  5. McAulay, R.J.; Malpass, M.L. Speech enhancement using a soft-decision noise suppression filter. IEEE Trans. Acoust. Speech Signal Process 1980, 28, 37–145. [Google Scholar] [CrossRef]
  6. Dendrinos, M.; Bakamides, S.; Carayannis, G. Speech enhancement from noise: A regenerative approach. Speech Commun. 1991, 10, 45–57. [Google Scholar] [CrossRef]
  7. Ephraim, Y.; Van Trees, H.L. A signal subspace approach for speech enhancement. In Proceedings of the IEEE International Conference on Acoustic, Speech, and Signal Processing, Minneapolis, MN, USA, 27–30 April 1993; pp. 355–358. [Google Scholar]
  8. Grais, E.M.; Sen, M.U.; Erdogan, H. Deep neural networks for single channel source separation. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Florence, Italy, 4–9 May 2014; pp. 3734–3738. [Google Scholar]
  9. Xu, Y.; Du, J.; Dai, L.-R.; Lee, C.-H. An Experimental Study on Speech Enhancement Based on Deep Neural Networks. IEEE Signal Processing Lett. 2013, 21, 65–68. [Google Scholar] [CrossRef]
  10. Liu, Q.; Wang, W.; Jackson, P.B.; Tang, Y. A perceptually-weighted deep neural network for monaural speech enhancement in various background noise conditions. In Proceedings of the 25th European Signal Processing Conference, Kos, Greece, 28 August–2 September 2017; pp. 1270–1274. [Google Scholar]
  11. Jarrett, D.P. Theory and Applications of Spherical Microphone Array Processing; Springer: New York, NY, USA, 2017. [Google Scholar]
  12. Benesty, J.; Chen, J.; Huang, Y. Microphone Array Signal Processing; Springer: New York, NY, USA, 2010. [Google Scholar]
  13. Zhao, Q.; Guo, F.; Zu, X.; Chang, Y.; Li, B.; Yuan, X. An Acoustic Signal Enhancement Method Based on Independent Vector Analysis for Moving Target Classification in the Wild. Sensors 2017, 17, 2224. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Nordhausen, K.; Oja, H. Independent component analysis: A statistical perspective. Wires Comput. Stat. 2018, 10, e1440. [Google Scholar] [CrossRef]
  15. Addisson, S.; Luis, V. Independent Component Analysis (ICA): Algorithm, Applications and Ambiguities; Nova Science Publishers: Hauppauge, NY, USA, 2018. [Google Scholar]
  16. Makino, S.; Lee, T.W.; Sawada, H. (Eds.) Blind Speech Separation; Springer: New York, NY, USA, 2007. [Google Scholar]
  17. Taseska, M.; Habets, E.A.P. Blind Source Separation of Moving Sources Using Sparsity-Based Source Detection and Tracking. IEEE/ACM Trans. Audio Speech Lang. Process. 2018, 26, 657–670. [Google Scholar] [CrossRef]
  18. Dekens, T.; Verhelst, W.; Capman, F.; Beaugendre, F. Improved speech recognition in noisy environments by using a throat microphone for accurate voicing detection. In Proceedings of the 18th European Signal Processing Conference, Aalborg, Denmark, 23–27 August 2010; pp. 1978–1982. [Google Scholar]
  19. Eisemann, E.; Durand, F. Flash photography enhancement via intrinsic relighting. ACM Trans. Graph. 2004, 23, 673–678. [Google Scholar] [CrossRef]
  20. Petschnigg, G.; Agrawala, M.; Hoppe, H.; Szeliski, R.; Cohen, M.; Toyama, K. Digital photography with flash and no-flash image pairs. ACM Trans. Graph. 2004, 23, 664–672. [Google Scholar] [CrossRef]
  21. Wang, S.; Liu, Z.; Lv, S.; Lv, Y.; Wu, G.; Peng, P.; Chen, F.; Wang, X. A Natural Visible and Infrared Facial Expression Database for Expression Recognition and Emotion Inference. IEEE Trans. Multimed. 2010, 12, 682–691. [Google Scholar] [CrossRef]
  22. John, V.; Tsuchizawa, S.; Mita, S. Fusion of thermal and visible cameras for the application of pedestrian detection. Signal Image Video Process. 2017, 11, 517–524. [Google Scholar] [CrossRef]
  23. Fendri, E.; Boukhriss, R.R.; Hammami, M. Fusion of thermal infrared and visible spectra for robust moving object detection. Pattern Anal. Appl. 2017, 20, 907–926. [Google Scholar] [CrossRef]
  24. Rickard, S.; Yilmaz, O. On the approximate w-disjoint orthogonality of speech. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, Orland, CA, USA, 13–17 May 2002; pp. 529–532. [Google Scholar]
  25. Ihara, T.; Handa, M.; Nagai, T.; Kurematsu, A. Multi-channel speech separation and localization by frequency assignment. IEICE Trans Fundam. 2003, J86-A, 998–1009. [Google Scholar]
  26. Aoki, M.; Yamaguchi, Y.; Furuya, K.; Kataoka, A. Modifying SAFIA: Separation of the target source close to the microphones and noise sources far from the microphones. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 2005, J88-A, 468–479. [Google Scholar]
  27. Sound Effect Lab. Available online: https://soundeffect-lab.info/sound/environment/ (accessed on 23 December 2021).
  28. Hashimoto Tech. Available online: https://hashimoto-tech.jp/local/advan/signwav (accessed on 23 December 2021).
  29. Fukui, M.; Shimauchi, S.; Hioka, Y.; Nakagawa, A.; Haneda, Y.; Ohmuro, H.; Kataoka, A. Noise-power estimation based on ratio of stationary noise to input signal for noise reduction. J. Signal Processing 2014, 18, 17–28. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Overview of the proposed method.
Figure 1. Overview of the proposed method.
Sensors 22 04473 g001
Figure 2. Waveform (Intersection noise).
Figure 2. Waveform (Intersection noise).
Sensors 22 04473 g002
Figure 3. Spectrogram (Intersection noise).
Figure 3. Spectrogram (Intersection noise).
Sensors 22 04473 g003
Figure 4. Waveform (White noise (0 dB)).
Figure 4. Waveform (White noise (0 dB)).
Sensors 22 04473 g004
Figure 5. Spectrogram (White noise (0 dB)).
Figure 5. Spectrogram (White noise (0 dB)).
Sensors 22 04473 g005
Figure 6. Waveform (White noise (−15 dB)).
Figure 6. Waveform (White noise (−15 dB)).
Sensors 22 04473 g006
Figure 7. Spectrogram (White noise (−15 dB)).
Figure 7. Spectrogram (White noise (−15 dB)).
Sensors 22 04473 g007
Table 1. Experimental condition.
Table 1. Experimental condition.
Target SpeechMale Voice
Noise signalIntersection noise,
white noise (0, −15 [dB])
Noise reduction threshold−90 to −30 [dB]
(in 10 [dB] steps)
Table 2. Intersection noise.
Table 2. Intersection noise.
Threshold (dB)SDRSIR
−900.704−3.347
−802.6320.008068
−708.3847.795
−609.59312.97
−507.39315.50
−404.35117.10
−301.52720.83
Table 3. White noise (0 dB).
Table 3. White noise (0 dB).
Threshold (dB)SDRSIR
−901.665−9.767
−803.199−7.361
−705.007−4.756
−606.838−1.369
−506.6462.083
−404.2365.033
−301.5228.821
Table 4. White noise (−15 dB).
Table 4. White noise (−15 dB).
Threshold (dB)SDRSIR
−907.5102.233
−809.5794.639
−7010.667.244
−609.76410.63
−507.53414.08
−404.41517.03
−301.54020.82
Table 5. SNR results.
Table 5. SNR results.
Noise SpeechSNR
Intersection noise−4.158
White noise (0 dB)−22.96
White noise (−15 dB)−10.96
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kawaguchi, J.; Matsumoto, M. Noise Reduction Combining a General Microphone and a Throat Microphone. Sensors 2022, 22, 4473. https://doi.org/10.3390/s22124473

AMA Style

Kawaguchi J, Matsumoto M. Noise Reduction Combining a General Microphone and a Throat Microphone. Sensors. 2022; 22(12):4473. https://doi.org/10.3390/s22124473

Chicago/Turabian Style

Kawaguchi, Junki, and Mitsuharu Matsumoto. 2022. "Noise Reduction Combining a General Microphone and a Throat Microphone" Sensors 22, no. 12: 4473. https://doi.org/10.3390/s22124473

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop