Next Article in Journal
BCA-MVSNet: Integrating BIFPN and CA for Enhanced Detail Texture in Multi-View Stereo Reconstruction
Previous Article in Journal
Story Forge: A Card-Based Framework for AI-Assisted Interactive Storytelling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Ultrasonic Jamming via Time–Frequency Mosaic for Anti-Eavesdropping Systems

by
Zichuan Yu
1,
Lu Tang
1,*,
Kai Wang
1,
Xusheng Tang
2 and
Hongyu Ge
1
1
School of Information Science and Engineering, Southeast University, Nanjing 211189, China
2
School of Cyber Science and Engineering, Southeast University, Nanjing 211189, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(15), 2960; https://doi.org/10.3390/electronics14152960
Submission received: 21 June 2025 / Revised: 16 July 2025 / Accepted: 23 July 2025 / Published: 24 July 2025
(This article belongs to the Topic Addressing Security Issues Related to Modern Software)

Abstract

To combat microphone eavesdropping on devices like smartphones, ultrasonic-based methods offer promise due to human inaudibility and microphone nonlinearity. However, existing systems suffer from low jamming efficiency, poor energy utilization, and weak robustness. Based on these problems, this paper proposes a novel ultrasonic-based jamming algorithm called the Time–Frequency Mosaic (TFM) technique, which can be used for anti-eavesdropping. The proposed TFM technique can generate short-time, frequency-coded jamming signals according to the voice frequency characteristics of different speakers, thereby achieving targeted and efficient jamming. A jamming prototype using the Time–Frequency Mosaic technique was developed and tested in various scenarios. The test results show that when the signal-to-noise ratio (SNR) is lower than 0 dB, the text Word Error Rate (WER) of the proposed method is basically over 60%; when the SNR is 0 dB, the WER of the algorithm in this paper is on average more than 20% higher than that of current jamming algorithms. In addition, when the jamming system maintains the same distance from the recording device, the algorithm in this paper has higher energy utilization efficiency compared with existing algorithms. Experiments prove that in most cases, the proposed algorithm has a better jamming effect, higher energy utilization efficiency, and stronger robustness.

1. Introduction

With the rapid development of recording technology and the widespread proliferation of electronic devices equipped with microphones, people are able to conveniently conduct voice calls and audio recordings. However, they are also confronted with the risk of being eavesdropped. Hackers exploit disguised recording devices or voice assistants found in smartphones to eavesdrop on private conversations and meetings [1,2,3], resulting in the leakage of user privacy. Covert microphones play a significant role in these espionage activities [4], which has drawn attention to anti-eavesdropping security.
Traditional anti-eavesdropping techniques primarily include audible acoustic interference and electromagnetic interference [5]. The audible acoustic interference method employs white noise generators [6] or high-frequency noise sources [7] to mask acoustic information. Although highly effective, this approach significantly compromises the quality of normal communication. Electromagnetic interference countermeasures operate by introducing electromagnetic signals to disrupt microphone sensor systems [8,9]. However, this method has high power consumption requirements, and many current microphone recording devices are equipped with built-in anti-interference algorithms, which results in a poor interference effect of anti-eavesdropping interference signals [5]. In addition, it may violate electromagnetic spectrum management regulations, creating legal risks [10].
Since ultrasonic sounds are inaudible to the human ear, they can be covertly utilized for anti-eavesdropping in smart devices [5], leading to the proposal of ultrasonic microphone jammers [11]. The inherent nonlinear characteristics of microphone systems enable the leakage of ultrasonic energy into the audible frequency band [12,13], thereby pioneering new avenues for covert acoustic interference and establishing an emerging research direction in information security [14,15,16]. Compared with traditional audible-band jamming methods, ultrasonic interference demonstrates distinctive advantages: (1) The inaudibility of high-frequency carriers ensures user conversational comfort; (2) directional transmission characteristics enable precise spatial coverage; (3) near-field energy concentration enhances power utilization efficiency.
Current research on anti-eavesdropping technologies based on ultrasonic interference has primarily developed along three technical pathways. Firstly, the ultrasonic carrier modulation technique leverages nonlinear intermodulation between ultrasonic waves and low-frequency noise to generate audible-band interference in microphone circuits [17]. Building on this, Chen et al. [18] designed a portable ultrasonic anti-eavesdropping gadget, while Shen et al. [19] developed a directional ultrasonic coverage system. Secondly, the high-power single-tone jamming technique induces microphone circuit saturation through high-intensity single-frequency ultrasonic waves. In this context, Gui et al. [20] proposed an audio amplitude attenuation system. Thirdly, the speech feature adaptation technique dynamically generates ultrasonic adversarial signals aligned with the temporal–spectral characteristics of speech. For instance, Reference [21] introduced an interference method based on human vocal feature matching to enhance jamming efficiency. Additionally, Reference [22] presented an auditory information masking technique that reduces speech recognition accuracy by minimizing the perceptual gap between noise and target speech.
Based on these three technical pathways, some anti-eavesdropping systems that utilize ultrasonic waves for interference have been proposed. The literature [23] presents an anti-eavesdropping system capable of resisting deep learning denoising and feature denoising. It addresses the security vulnerability in traditional ultrasonic interference, where 75% of content can be recovered by using adaptive gain compensation and broad-band coverage, but it does not consider the design of system power consumption and cost. The literature [5] converts random noise into ultrasonic signals through DSB modulation and generates full-band audible noise via self-demodulation, achieving progress in concealment and operating distance; however, it does not consider specific optimization for human voices. The literature [3] solves the pain point of indiscriminate interference in traditional schemes through a selective interference mechanism and triple anti-recovery security design, but it has residual low-frequency noise and fails to fundamentally eliminate the impact of nonlinear effects. The literature [1] proposes an evaluation framework and standards for ultrasonic anti-eavesdropping systems, reveals the general vulnerability of existing UMJ schemes, and puts forward optimization paths for the design of ultrasonic interference systems. Nevertheless, its framework has insufficient coverage of dynamic scenarios and certain limitations.
Based on the above analysis, existing anti-eavesdropping interference systems based on ultrasonic waves still face three major limitations: (1) Spectral Mismatch—Most methods generate uniform-spectrum interference, failing to account for the nonuniform energy distribution characteristics inherent to human speech [24,25]; (2) Temporal Blindness—Existing systems lack real-time adaptation to speech dynamics, resulting in energy wastage during speech pauses [22]; (3) Hardware Constraints—The limited conversion efficiency of ultrasonic transducers restricts practical deployment scenarios.
Therefore, how to design an anti-eavesdropping system with good jamming effect, high jamming efficiency, and strong robustness based on ultrasonic interference remains a problem worthy of attention. To address these issues, this work proposes the Time–Frequency Mosaic (TFM) jamming paradigm, fundamentally redefining the design of ultrasonic anti-eavesdropping systems. The core innovations of this work include the following:
  • Time–Frequency Joint Interference: A novel joint jamming method that integrates temporal and frequency domain characteristics is proposed for the first time. Unlike typical noise injection techniques, TFM not only adjusts its frequency but also modulates its temporal sequence.
  • Adaptive Ultrasonic Encoding Strategy: Based on the state of the speech frame, the jamming strategy is dynamically switched to enhance the robustness of the jamming.
  • Energy-Constrained Hardware Co-Design: The amplitude of the jamming signal is dynamically adjusted according to the energy of the speech frame, ensuring effective jamming while reducing power consumption.
The remainder of this paper is organized as follows: Section 2 analyzes the principles of the anti-eavesdropping system; Section 3 introduces the principles of the proposed TFM technology; Section 4 constructs the hardware platform of the anti-eavesdropping system and conducts tests; Section 5 concludes the research.

2. System Principle and Analysis

2.1. Out-of-Band Signal Conversion in Microphone

The signal path of the speech acquisition hardware consists of microphones, amplifiers, low-pass filters, and analog-to-digital converters(ADCs), as shown in Figure 1a. And the frequency range for human hearing and speech communication usually covers 20 Hz to 20 kHz. As a result, most microphone systems operate within this range. Even if a microphone is capable of detecting frequencies above 20 kHz, a low-pass filter efficiently removes these high-frequency signals. The typical sampling rate of analog-to-digital converters is 44.1 kHz as per the Nyquist sampling theorem, which states that the digital signal frequency should be below 22 kHz. For recording systems, in-band audible speech signals up to 22 kHz and inaudible ultrasonic signals above 25 kHz are typical out-of-band signals.
Although electronics, such as amplifiers, are typically designed with a linear input–output relationship, the actual devices exhibit nonlinear characteristics. Due to the nonlinear characteristics, the electronics may suffer from intermodulation distortion (IMD), which results in the presence of frequency components in the output that are not present in the input signal. The nonlinear nature of the microphone can be approximated as the transfer function equation of the input and output signals having a nonlinear quadratic square term. When the input signal is not of a single frequency, the quadratic term in the nonlinear equation generates harmonics and cross-products of the input signal, which cause the output signal to contain new frequencies. Therefore, as shown in Figure 1b,c, the inherent nonlinear properties of the device can be utilized to achieve the effect of converting out-of-band inaccessible signals into in-band audible signals within the voice recording system by designing suitable ultrasonic input signals.
For two signals a and b with multiple frequency components, the nonlinear components s n l t generated by the system can be expressed as
s n l t = i j β A a , i A b , j cos 2 π f a , i ± f b , j t
where β is the nonlinearity coefficient, A a , i represents the amplitudes of the i-th component in signal a, A b , j represents the j-th in signal b, f a , i and f b , j represent the i-th frequency component of a and the j-th frequency component of b, respectively.

2.2. Eavesdropping Models Analysis

Consider the scenario where the eavesdropper’s objective is to record private conversations in a closed and confidential setting, such as small conference rooms or private offices. Eavesdropping devices come in various forms, including ordinary smartphones or other covert recording devices. However, all of these devices have microphone recording systems at their core. The device for eavesdropping can be controlled by the eavesdropper either manually, by switching it on and off, or through software. Additionally, we assume that the eavesdropping device is carried by one of the interlocutors. It is probable that the device is hidden on the opponent’s body or at least in close proximity, such as being secretly placed under the table.
For the anti-eavesdropping scenario, it is assumed that the two parties are adults conversing at a normal volume in a relatively quiet environment. It is important to note that the general hearing range of the adult ear is below 18 kHz; therefore, 25 kHz ultrasonic sound should be inaudible to the interlocutor. As shown in Figure 2, users wishing to counter eavesdrop can preemptively perform speech analysis on the jamming device and activate the jamming device when a conversation is required, and the jamming device emits jamming noise to the approximate location of the possible eavesdropping device.

3. Time–Frequency Mosaic Interference

Based on the time–frequency characteristics of speech signals, this work proposes a TFM interference strategy. By jointly attacking the critical time–frequency features of speech signals through adaptive mosaic-style coverage of key speech components, this method achieves precise interference in both temporal and spectral domains to effectively disrupt Automatic Speech Recognition (ASR) systems while maintaining imperceptibility to human auditory perception. The algorithm operates in three stages:
  • Real-time Speech Hybrid Feature Analysis;
  • Dynamic Interference Signal Synthesis;
  • Energy-Constrained Synthesis and Robustness Enhancement.
The general structure of the TFM interference system can be observed in Figure 3 and comprehensively features two core elements: the analysis module and the hardware module. The analysis segment functions primarily to formulate interference signals, while the hardware segment produces the necessary output signals based on the interference library advanced by the analysis module. The ultrasound transducer array subsequently transmits these signals.

3.1. Real-Time Speech Hybrid Feature Analysis

TFM first conducts a mixed feature analysis on the voice signal.
Continuous speech constitutes a nonstationary signal, necessitating its decomposition into quasi-stationary segments to facilitate localized feature analysis. For a speech signal x ( n ) with sampling rate f s , the k-th frame signal x k ( n ) is defined as
x k ( n ) = x ( n + k M ) , 0 n < N
Here, N = 0.02 f s (20 ms frame length), and M = 0.5 N (50% overlap).
A Hamming window is applied to the framed quasi-stationary segments during the windowing process to suppress spectral leakage. The Hamming window function w ( n ) is given in (3), and the Hamming-window signal x k w i n ( n ) is given in (4).
w ( n ) = 0.54 0.46 cos 2 π n N 1
x k w i n ( n ) = x k ( n ) · w ( n )
Subsequently, the Short-Time Fourier Transform (STFT) is computed for each windowed frame to obtain time–frequency representations. For the k-th frame X k ( m ) , its STFT is defined as
X k ( m ) = n = 0 N 1 x k w i n ( n ) e j 2 π m n / N , 0 m < N
Then, the formant structure is extracted using a 12th-order linear predictive coding (LPC) model, and its power spectral density formula P LPC ( f ) is shown in (6) [26,27].
P LPC ( f ) = σ 2 1 i = 1 12 a i e j 2 π f i 2
Here, σ 2 is the noise variance, and a i represents the LPC coefficients optimized through autocorrelation minimization.
The dual-spectral analysis approach is adopted to facilitate hybrid feature extraction, where STFT provides coarse-grained time–frequency localization, while LPC precisely identifies ASR-dependent formant regions.

3.2. Dynamic Interference Signal Synthesis

Based on feature analysis, TFM will generate dynamic jamming signals.
The concentration of speech information and energy within critical frequency components (e.g., fundamental frequency, formants) necessitates localized feature targeting based on spectral analysis outcomes. A differentiated interference strategy is then developed to maximally disrupt ASR-discriminative features.
First, perform mel scale energy focusing by converting the STFT magnitudes into the mel scale using 40 triangular filters H m ( k ) . The mel scale emulates the human ear’s nonlinear auditory characteristics and covers the primary speech frequency range (300 Hz–4 kHz). The top 5% of mel bands with the highest energy are selected for prioritized interference. This energy-concentrated band screening avoids resource waste from full-band interference. H m ( k ) is shown in (7) [28].
H m ( k ) = 0 k < f ( m 1 ) k f ( m 1 ) f ( m ) f ( m 1 ) f ( m 1 ) k f ( m ) f ( m + 1 ) k f ( m + 1 ) f ( m ) f ( m ) < k f ( m + 1 ) 0 k > f ( m + 1 )
And mel scale f mel is given by (8) [29,30]
f mel = 2595 log 10 1 + f 700
where
f = 700 ( 10 m / 2595 1 )
Perform formant neighborhood detection to precisely disrupt the ASR-dependent formant features while reducing the interference bandwidth. The formant bandwidth typically ranges from 50 to 100 Hz. A Gaussian window is applied to cover their energy diffusion range. The first three formant frequencies F1, F2, and F3 are extracted from P L P C ( f ) , defining the interference frequency bands B i as
B i = [ F i Δ f , F i + Δ f ] , Δ f = 50 Hz , i = 1 , 2 , 3
Based on the selected interference bands, compute the psychoacoustic masking thresholds. The masking threshold T b shown in (11) is calculated within 24 critical bands of the Bark scale, where E b denotes the energy of band b, and S b represents the predefined masking offset specified in the ISO/IEC 11172-3 standard [31].
T b = 10 log 10 ( E b ) S b
After obtaining the interference frequency bands and masking values, the interference signal can be generated. This work designs two strategies for interference signal generation: low-frequency phase-inverted noise interference and mid-to-high-frequency swept-frequency interference. The fundamental frequency and harmonic structure are core features of speech periodicity, and phase-inverted interference can disrupt the temporal waveform correlation. In the mid-to-high-frequency range, formant frequencies vary over time (e.g., vowel transitions). Swept-frequency signals provide dynamic coverage, while noise introduces randomness to prevent ASR systems from recovering speech content by tracking formant trajectories.
A phase-inverted signal s low ( t ) shown in (12) is injected into the formant regions below 500 Hz to disrupt the fundamental frequency harmonic structure, where n l o w ( t ) represents band-limited Gaussian noise with a bandwidth B = 100 Hz.
s low ( t ) = A · sin ( 2 π F 0 t + π ) + n low ( t )
A linear chirp signal s sweep ( t ) shown in (13) is generated in the mid-to-high-frequency bands to mask time-varying formants, with T = 20 ms, and F e n d = F i + 100 Hz.
s sweep ( t ) = A · sin 2 π F i t + β 2 t 2 , β = F end F i T

3.3. Energy-Constrained Synthesis and Robustness Enhancement

This section will introduce the energy constraint mechanism and robustness enhancement mechanism in the TFM so as to improve the energy utilization efficiency and robustness of the overall system.
Due to significant fluctuations in speech energy (e.g., abrupt increases in plosive sounds), real-time adjustment of interference intensity is required to prevent insufficient interference in low-energy frames and overload distortion in high-energy frames. Dynamic amplitude control enforces energy constraints by adjusting the interference amplitude A b in each sub-band B i to satisfy
A b = min η b , 0.8 · E b
Hardware experiments indicate that interference exceeding 80% of the original signal’s amplitude tends to cause temporal waveform distortion (e.g., clipping). Therefore, dual constraints are implemented to balance stealthiness and interference intensity. Here, η b ( T b ) represents the masking threshold, and the coefficient 0.8 is experimentally optimized to ensure signal stability.
The time-domain interference signal s ( n ) shown in (15) is reconstructed through inverse STFT and the Overlap-Add (OLA) method. Each s k corresponds to the slow (low-frequency) and sweep (mid–high frequency) components of different frequency bands. The OLA method ensures inter-frame signal continuity, prevents “click” artifacts, enhances the naturalness of the interference signal, and reduces human auditory sensitivity.
s ( n ) = k s k ( n k M ) · w ( n k M )
After completing the interference signal synthesis, robustness enhancement is primarily achieved through voiced/unvoiced state-adaptive control and formant tracking smoothing, ensuring sustained effectiveness of interference signals against dynamic speech.
Firstly, voiced/unvoiced detection is performed to distinguish speech frames as voiced sounds (periodic vocal fold vibrations) or unvoiced sounds (random noise characteristics), enabling dynamic adjustment of interference strategies. Unvoiced sounds (e.g., /s/, /t/) exhibit low-energy and high-frequency concentration, necessitating a switch to broad-band noise interference to avoid failure of narrow-band interference designed for voiced sounds in unvoiced segments.
Voiced/unvoiced detection utilizes short-term energy E and zero-crossing rate Z C R to classify speech frames as follows:
E = 1 N n = 0 N 1 x k 2 ( n )
Z C R = 1 2 N n = 1 N 1 | sgn ( x k ( n ) ) sgn ( x k ( n 1 ) ) |
The specific decision strategy is
V o i c e d f r a m e E > θ E , Z C R < θ Z
where θ E = 0.3 m a x ( E ) , θ Z = 0.1 .
Following detection, dynamic strategy adjustments are implemented.
For voiced frames: Apply low-frequency formant-enhanced interference targeting the fundamental frequency F0 and harmonic components by increasing the phase-inverted noise amplitude Alow, simultaneously, to suppress high-frequency interference to reduce energy allocation of swept-frequency signals in mid–high frequencies and avoid spectral overload.
For unvoiced frames: Inject mid-to-high-frequency noise in fricative/plosive regions (e.g., 4–8 kHz) by augmenting broad-band noise interference. Concurrently, attenuate low-frequency interference to avoid energy waste in base-band unvoiced segments lacking fundamental frequency components.
Moreover, formant frequencies vary smoothly between adjacent frames (e.g., during sustained vowel segments), requiring smoothing processing to avoid abrupt jumps in interference signals and enhance the obfuscation effect. Specifically, exponential smoothing is applied to the formant frequencies to suppress inter-frame abrupt changes in formant frequencies, thereby improving the continuity of interference signals.
F ^ i ( k ) = α F ^ i ( k 1 ) + ( 1 α ) F i ( k ) , α = 0.9
The formants F i ( k ) in (19) extracted from the k-th frame are smoothed with α = 0.9, achieving a relative balance between tracking speed and stability. This prevents discontinuities in the interference band B i shown in (20) caused by abrupt formant transitions, ensuring gradual variation of the frequency modulation (FM) slope β in the swept-frequency signal s s w e e p ( t ) .
B i ( k ) = [ F ^ i ( k ) Δ f , F ^ i ( k ) + Δ f ]
Based on the smoothed formant frequencies F ^ i , recalculate the interference frequency bands and adjust the swept-frequency signal parameters F i and F e n d to ensure that the interference consistently covers the smoothed target frequency bands.
Robustness enhancement achieves dynamic interference optimization through two core technologies: These include voiced/unvoiced state-adaptive control and formant tracking smoothing. By dynamically allocating interference band energy based on voiced/unvoiced classification—for instance, focusing on high-frequency interference instead of low-frequency formant coverage when plosive sounds like /t/ occur—stealthiness is improved. Simultaneously, exponential smoothing filtering eliminates inter-frame jumps in formant tracking, ensuring time-frequency continuity of interference signals. Additionally, the voiced/unvoiced adaptive strategy reduces the total energy of interference signals by approximately 18%, optimizing energy efficiency while maintaining attack effectiveness. The synergistic interaction of these two technologies enables the synthesized interference signals to adapt to dynamic speech characteristics, resist the short-term spectral recovery capabilities of ASR systems, and ultimately enhance the robustness and stealthiness of adversarial attacks.

4. Evaluation and Discussion

In this section, a physical platform was built based on the proposed TFM, and the TFM was evaluated in terms of energy utilization efficiency, jamming effect, robustness, and performance in real environments to verify its performance.

4.1. Experimental Set Up

The hardware circuit implementation of the Time–Frequency Mosaic interference system is displayed in Figure 4. The system mainly includes the power module, the interference signal generation and processing module, and the ultrasonic transmission array. The interference signal generation and processing module consists of FPGA ZYNQ7020 (Produced by MicroPhase, Shanghai, China), digital-to-analog converter DAC8554 (Produced by Mindesigner, Changsha, China), low-pass filters, and two power amplifiers. For optimum interference effectiveness, the ultrasonic transmission array has been fitted with probes that are differentially distributed. Each probe transmits two distinctive Time–Frequency Mosaic interference signals.
Figure 5 is a simplified schematic of the experiment. Testers for the experiment were preregistered on the TFM system. During the experiment, the ultrasound array was directed toward the recording device. First, a recording was performed without interference. Then, the ultrasound jamming device was turned on, and the recording device was jammed with white noise and the TFM, respectively. We used a Redmi Note 11T Pro smartphone as the recording device.
Evaluating the interference effect of speech interference algorithms typically involves the use of a series of objective and subjective assessment metrics to ensure the performance and practical usability of the interference algorithms. The Mean Opinion Score (MOS) method is a subjective evaluation approach involving experts (listeners) who rate speech. Testing the interfered speech with a Speech-To-Text System (STT) and obtaining a transcribed text, followed by a comparison with the correct text, allows for the calculation of the Word Error Rate (WER) [32]. To reduce the subjectivity of the evaluation, this paper employed the STT testing method, using the WER to indicate the quality of interference. A higher WER suggests poorer recognition by the STT system, which in turn indicates a better interference effect.

4.2. Interference Energy Utilization Efficiency Analysis

The recording device was positioned at a consistent distance of 20 cm from the interfering device. The human voice was recorded both with and without interference from white noise and the TFM. The speaker and spoken text remained constant throughout the recordings. The resulting audio files were exported and analyzed.
To compare the effects of interference between TFM noise and white noise on the same sound, we plotted the time domain and time–frequency domain waveforms of the pure sound, the sound interfered with by white noise, and the sound interfered with by the TFM in Figure 6. The power of white noise is uniformly dispersed throughout the time–frequency domain, whereas TFM noise can be targeted to cover the specific composition of the human voice to achieve a better masking effect.
Meanwhile, it can be observed from Figure 6 that the energy of the TFM appears to be higher than that of white noise at the same distance. To confirm the energy transfer efficiency of the interference signal, we recorded the white noise interference signal and the TFM interference signal at the same distance. We calculated the SNR of the two recorded interference noises with the same segment of pure human voice signal distribution. We then obtained the SNR under different distances, as shown in Figure 7. The distance refers to the distance from the ultrasonic array to the sound-receiving device. Figure 7 shows that the noise generated by the TFM interference signal is higher than that of the white noise interference signal in the low-distance case by 2 dB. In the long-distance case, the two SNR curves rapidly increase and converge due to the limited power of the transmitting equipment and ultrasonic propagation in the air with loss, resulting in a rapid reduction of the noise generated by the interference signal. Therefore, the TFM jamming signal requires less power than white noise to achieve the same signal-to-noise ratio, demonstrating the superior energy efficiency of the TFM jamming system.

4.3. Interference Effect Analysis

To comprehensively evaluate the interference effect, we conducted tests using four different STT systems: two major commercial STT systems, Tencent STT and Xunfei STT; a free STT system, Diyun STT; and a commonly used open-source STT system, DeepSpeech [33].
Tencent STT boasts core strengths in high real-time performance and engineering capabilities. Its low-latency feature makes it the first choice for scenarios such as real-time subtitles in Tencent Meetings and voice input in WeChat. iFLYTEK Hear (Xunfei STT), developed by iFLYTEK, excels in professional field recognition and becomes a benchmark solution in the judicial and medical industries. Diyun AI Speech (Diyun STT), as a representative of embedded speech recognition, is a lightweight model that is compatible with Internet of Things (IoT) chips, meeting the stringent requirements of industrial inspection and smart home with ultra-low-power consumption. DeepSpeech, as an open-source project by Mozilla, supports cross-language recognition of over 80 languages, providing a a flexible open-source foundation for academic research and multilingual customization.
Note that the TFM produces higher power noise at the same distance. To fairly compare the interference effect of the noise produced by the interfering signal, we extracted the noise and adjusted the power ratio of the two types of noise to one. Then, we mixed the pure speech signals based on Mandarin with the noise signals in the digital domain and fed them into the STT system for comparison.
A comparison was performed between TFM interference and the jamming method in Backdoor [17], where the noise parameters replicated Backdoor’s design: band-limited white noise (0–12 kHz) was amplitude-modulated onto a 40 kHz ultrasonic carrier to exploit microphone nonlinearity. Under various SNRs, the recordings with TFM interference and white noise interference were input into the STT systems. The result is shown in Figure 8.
Figure 8 illustrates the WER values of each STT system within the SNR range of [−5, 5] dB. Among them, the red line represents the change in WER of the proposed STT with SNR, the white line represents the change in WER of white noise with SNR, and the black line represents the change in WER without any interference with SNR.
Figure 8 shows that our TFM interference design performed better than white noise interference, particularly when the SNR was below 0 dB. When white noise was present at an SNR of −5 dB, Tencent STT and Xunfei STT produced a much lower WERs compared to Diyun STT and Deep Speech. Additionally, the interference effects of other signal-to-noise ratios of white noise were weaker for Tencent STT and Xunfei STT. This suggests that Tencent STT and Xunfei STT have anti-interference designs for white noise, which greatly attenuate the white noise interference effect.
At an SNR of 0 dB, the average WER for TFM interference was 20% higher than for white noise. For the two main commercial STT systems, there was a greater difference in interference performance between TFM interference and white noise interference at low SNRs than in open-source STT systems. This variation can be associated with the use of noise reduction algorithms in commercial STT systems, showing the weaker suppression of TFM interference by current commercial noise reduction methods. This illustrates the resilience of TFM interference.
It should be noted that Diyun STT had a similar white noise interference effect to TFM at a low SNR. However, its WER dropped rapidly after the SNR was above 0 dB, indicating that the white noise mainly suppressed Diyun STT due to the lack of internal speech noise reduction and speech recovery extraction capabilities in this STT system.
In order to better illustrate the effect of TFM, we used the same equipment to transmit different interference signals, namely, noise in TFM, Backdoor [17], and noise in Patronus [34], and tested the actual interference effect in 20 cm, 40 cm, and 60 cm. The WER results of the three methods at different distances are shown in Table 1.
From Table 1, it can be found that TFM had the best interference effect, while white noise was the worst. At a distance of 20 cm, TFM was 43.8% better than Patronus; at a distance of 60 cm, TFM was 157.5% better than Patronus, which proves that the interference effect of TFM attenuates the least with distance. Meanwhile, TFM can generate higher-power interference in the interfered devices under the condition that the transmitting devices have the same power. This reflects that TFM interference has high energy utilization efficiency.

4.4. Robustness Evaluation of Speech Enhancement

In addition to testing the speech interference effect of the TFM system, the robustness of the TFM method needs to be evaluated. Generally speaking, the anti-eavesdropping interference signal may be denoised by the hardware filtering built-in in the recording equipment and the speech enhancement algorithm of the internal speech processing chip, which may weaken the effect of the anti-eavesdropping interference system.
Given that the 0–8 kHz filter has a higher power to filter out white noise, and considering the signal-to-noise ratio experiments conducted in the previous section with the same-distance recording device, the power of the white noise delivered to the smart device is relatively low compared to the TFM. As previously done, we adjusted the noise generated by the two interferences to the same power. Then, we mixed it with pure speech in the digital domain to obtain the same power of the speech signal masked by the different interfering signals.
The SOTA algorithm [35] was used to enhance the mixed speech signal, which was then input into the STT system. Figure 9 displays the recognition effect of the speech-enhanced interfering signals by each STT system with different signal-to-noise ratios.
The results indicate that the recognition rates of Xunfei STT and Deepspeech STT for speech-enhanced mixed speech remained unchanged, and the effect of speech enhancement was limited. The recognition rate of Tencent STT for white noise remained unchanged.
However, the WER for TFM increased. This may be due to a conflict between the speech enhancement algorithm used in Tencent STT and the TFM interference after the SOTA algorithm, resulting in an amplification of the interference effect. Xunfei STT, Deepspeech, and Tecent STT have built-in speech enhancement algorithms, which means that the system’s speech recognition accuracy remains relatively unchanged even after being enhanced by the SOTA algorithm. However, Diyun STT, which lacks a built-in speech enhancement algorithm, experienced a significant change in recognition accuracy.
Figure 9 demonstrates that Diyun STT produced a significant improvement in the speech recognition rate after speech enhancement, reducing the WER under white noise by an average of 25%. This suggests that the speech enhancement algorithm greatly weakens the effect of white noise interference on Diyun STT, indicating poor robustness of the white noise interference signal. Compared to white noise interference, the error rate of text recognition for TFM-interfered speech was only reduced by about 5% on average after speech enhancement by the SOTA algorithm. Despite this improvement, the overall interference efficiency of TFM interference on speech remained high, which makes the Diyun STT robust to speech enhancement. Additionally, the noise generated by the TFM interference signal showed robustness to the SOTA speech enhancement algorithm.

4.5. Testing in Real-World Scenarios

Previous experiments were conducted to jam a single speaker. To verify the universality of the jamming system, multiple speaker scenarios were considered. Five different speakers were selected and allowed to have a conversation while following the same text. The audio was recorded and jammed using an ultrasonic eavesdropping prevention system. After adjusting the SNR, the audio was fed into four STT systems. The results presented in Figure 10 indicate that the interference effect of white noise remained consistent in the case of multiple speakers. However, the interference effect of TFM decreased while still maintaining a good level of interference. This is due to the fact that the speech of different individuals has varying acoustic characteristics, such as fundamental frequency and resonance peak frequency. Additionally, the interference band of the TFM signals was limited to achieve exceptional time–frequency domain coverage, which led to a degradation of the interference effect.
In the previous multispeaker test, our experimenters were all male. To ensure objectivity, future tests will include experimenters of different genders. In order to better evaluate the performance of TFM interference during real scenarios, we selected Xunfei STT and Tecent STT. These commercial STT systems have better stability and universality, making them closer to real scenarios. Keeping the SNR at 0 dB, we tested the interference effect on three groups: single male, single female, and mixed. Each group consisted of five interlocutors who followed the same conversation script. Interference was introduced using white noise and TFM, and the average speech recognition results are shown in Figure 11. From Figure 11, we can find that the interference effect of white noise in the different genders basically did not change much, and the interference effect on single women was relatively weak. In terms of TFM interference, there was a more pronounced difference in the interference effect between genders, with the best result being 65.2% for women and the worst being 57.4% for mixed genders.
Regarding the experimental results above, it is suggested that the concentration of frequency in women’s voices and their faster speech speed may have contributed to the observed differences. Women’s voices have a relatively high resonance peak frequency and a narrower bandwidth, while men’s voices have lower fundamental and resonance peak frequencies, smaller spacing between overtones, and a larger bandwidth. Additionally, white noise may have a relatively poor effect on the interference of relatively high-frequency sounds with narrower bandwidths. TFM interference is designed by obtaining certain sound information in advance, making it specialized. Female voices have higher frequencies, larger spacing, and fewer main harmonics, simplifying the design of TFM interference signals and resulting in a stronger interference effect. Male voices, on the other hand, have a wider frequency bandwidth and more harmonics, making it difficult for TFM jamming to completely cover them, resulting in a weaker jamming effect. When both males and females are present in a group, the frequency at which they experience jamming increases, resulting in a more severe jamming effect.

5. Conclusions

This paper employed time–frequency analysis methods and proposed an efficient anti-eavesdropping system based on TFM interference, with its core contributions including the proposal of a novel time–frequency joint interference strategy that performs interference in both the time domain and frequency domain to enhance the system’s interference capability, dynamic ultrasonic encoding that switches interference strategies according to voice states to boost the system’s robustness, and hardware platform design and experiments where we have developed a low-power hardware co-design platform and designed experiments from aspects such as energy efficiency, the interference effect, robustness evaluation, and real-scenario testing, fully demonstrating its excellent interference effect, robustness, and energy efficiency.
Experimental findings showcase that TFM interference significantly disrupted STT systems at an SNR of 0 dB, yielding WERs exceeding 40%. This underscores the robustness against noise reduction algorithms in low-SNR scenarios. Moreover, TFM exhibited markedly superior interference efficiency compared to white noise in high-SNR conditions with open-source STT systems, highlighting its outstanding energy utilization efficiency. This study emphasizes TFM’s potential as a practical audio security solution for safeguarding privacy and information in an inter-connected world.

Author Contributions

Conceptualization and writing—original draft, Z.Y.; supervision, L.T.; supervision, K.W.; writing—review and editing, X.T.; writing—review and editing, H.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Open Fund of Key Laboratory of Short-Range Radio Equipment Testing and Evaluation, Ministry of Industry and Information Technology (6809010015).

Data Availability Statement

The authors confirm that the data supporting the findings of this study are available within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gao, M.; Chen, Y.; Li, Y.; Zhang, L.; Liu, J.; Lu, L.; Lin, F.; Han, J.; Ren, K. A Resilience Evaluation Framework on Ultrasonic Microphone Jammers. IEEE Trans. Mob. Comput. 2023, 23, 1914–1929. [Google Scholar] [CrossRef]
  2. Chen, Y.; Gao, M.; Li, Y.; Zhang, L.; Lu, L.; Lin, F.; Han, J.; Ren, K. Big brother is listening: An evaluation framework on ultrasonic microphone jammers. In Proceedings of the IEEE INFOCOM 2022—IEEE Conference on Computer Communications, Online, 2–5 May 2022; pp. 1119–1128. [Google Scholar]
  3. Chen, Y.; Gao, M.; Liu, Y.; Liu, J.; Xu, X.; Cheng, L.; Han, J. Implement of a secure selective ultrasonic microphone jammer. CCF Trans. Pervasive Comput. Interact. 2021, 3, 367–377. [Google Scholar] [CrossRef]
  4. Katz, I.R.; Mack, R.L.; Marks, L.; Rosson, M.B.; Nielsen, J. Human factors in computing systems. In Proceedings of the CHI ’95 Conference Companion: CHI ’95, Mosaic of Creativity, Conference on Human Factors in Computing Systems, Denver, CO, USA, 7–11 May 1995. [Google Scholar]
  5. Ma, X.; Song, Y.; Wang, Z.; Gao, S.; Xiao, B.; Hu, A. You can hear but you cannot record: Privacy protection by jamming audio recording. In Proceedings of the ICC 2021—IEEE International Conference on Communications, Montreal, QC, Canada, 14–23 June 2021; pp. 1–6. [Google Scholar]
  6. Colton, G.D. High-Tech Approaches to Breeching Examination Security. Espionage 101. In Proceedings of the Annual Meeting of the National Council on Measurement in Education, Chicago, IL, USA, 25–27 March 1997. audiotape recordings. [Google Scholar]
  7. Xu, D.; Yan, X.; Chen, B.; Yu, L. Energy-Constrained Confidentiality Fusion Estimation Against Eavesdroppers. IEEE Trans. Circuits Syst. II Express Briefs 2022, 69, 624–628. [Google Scholar] [CrossRef]
  8. Tung, Y.C.; Shin, K.G. Exploiting Sound Masking for Audio Privacy in Smartphones. In Proceedings of the the 2019 ACM Asia Conference, Beijing, China, 16–18 December 2019. [Google Scholar]
  9. Kune, D.F.; Backes, J.; Clark, S.S.; Kramer, D.; Reynolds, M.; Fu, K.; Kim, Y.; Xu, W. Ghost Talk: Mitigating EMI Signal Injection Attacks against Analog Sensors. In Proceedings of the 2013 IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 19–22 May 2013; pp. 145–159. [Google Scholar]
  10. Zhang, G.; Yan, C.; Ji, X.; Zhang, T.; Zhang, T.; Xu, W. DolphinAtack: Inaudible Voice Commands. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017. [Google Scholar]
  11. Chen, Y.; Li, H.; Nagels, S.; Li, Z.; Lopes, P.; Zhao, B.Y.; Zheng, H. Understanding the effectiveness of ultrasonic microphone jammer. arXiv 2019, arXiv:1904.08490. [Google Scholar]
  12. Abuelma’atti, M.T. Analysis of the effect of radio frequency interference on the DC performance of bipolar operational amplifiers. IEEE Trans. Electromagn. Compat. 2003, 45, 453–458. [Google Scholar] [CrossRef]
  13. Gago, J.; Balcells, J.; GonzÁlez, D.; Lamich, M.; Mon, J.; Santolaria, A. EMI susceptibility model of signal conditioning circuits based on operational amplifiers. IEEE Trans. Electromagn. Compat. 2007, 49, 849–859. [Google Scholar] [CrossRef]
  14. Yanjing, W.; Zhuoran, M.; Wenyuan, X. Research of anti-eavesdropping technology based on electromagnetic interference against analog sensors. Electron. Technol. 2016, 45, 47–51. [Google Scholar]
  15. Roy, N.; Shen, S.; Hassanieh, H.; Choudhury, R.R. Inaudible Voice Commands: The Long-Range Attack and Defense. In Proceedings of the Networked Systems Design and Implementation, Renton, WA, USA, 9–11 April 2018. [Google Scholar]
  16. Sun, K.; Chen, C.; Zhang, X. “Alexa, stop spying on me!” speech privacy protection against voice assistants. In Proceedings of the 18th Conference on Embedded Networked Sensor Systems, Virtual Event, 16–19 November 2020; pp. 298–311. [Google Scholar]
  17. Roy, N.; Hassanieh, H.; Roy Choudhury, R. Backdoor: Making microphones hear inaudible sounds. In Proceedings of the 15th Annual International Conference on Mobile Systems, Applications, and Services, Niagara Falls, NY, USA, 19–23 June 2017; pp. 2–14. [Google Scholar]
  18. Chen, Y.; Li, H.; Teng, S.Y.; Nagels, S.; Li, Z.; Lopes, P.; Zhao, B.Y.; Zheng, H. Wearable Microphone Jamming. In Proceedings of the CHI ’20: 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25–30 April 2020; pp. 1–12. [Google Scholar]
  19. Shen, H.; Zhang, W.; Fang, H.; Ma, Z.; Yu, N. JamSys: Coverage Optimization of a Microphone Jamming System Based on Ultrasounds. IEEE Access 2019, 7, 67483–67496. [Google Scholar] [CrossRef]
  20. Gui, H.; Yan, C.; Cheng, Y.; Ji, X.; Xu, W. Black Hole of Sound: Constructing an Anti-eavesdropping Area Using Ultrasound. In Proceedings of the 2022 IEEE 6th Conference on Energy Internet and Energy System Integration (EI2), Chengdu, China, 28–30 October 2022; pp. 1717–1722. [Google Scholar]
  21. Brungart, D.S.; Simpson, B.D.; Ericson, M.A.; Scott, K.R. Informational and energetic masking effects in the perception of multiple simultaneous talkers. J. Acoust. Soc. Am. 2001, 110, 2527–2538. [Google Scholar] [CrossRef] [PubMed]
  22. Huang, P.; Wei, Y.; Cheng, P.; Ba, Z.; Lu, L.; Lin, F.; Zhang, F.; Ren, K. InfoMasker: Preventing Eavesdropping Using Phoneme-Based Noise. In NDSS; Zhejiang University: Hangzhou, China, 2023. [Google Scholar]
  23. Han, Z.; Ma, J.; Xu, C.; Zhang, G. UltraJam: Ultrasonic adaptive jammer based on nonlinearity effect of microphone circuits. High-Confid. Comput. 2023, 3, 100129. [Google Scholar] [CrossRef]
  24. Fang, X.; Zhao, D.; Zhang, L. Unimodular Multi-Input Multi-Output Waveform and Mismatch Filter Design for Saturated Forward Jamming Suppression. Sensors 2024, 24, 5884. [Google Scholar] [CrossRef] [PubMed]
  25. Yu, L.J.T.; Zhao, R.; Ji, S.; Ngai, E.C.; Wu, C. USpeech: Ultrasound-Enhanced Speech with Minimal Human Effort via Cross-Modal Synthesis. arXiv 2024, arXiv:2410.22076. [Google Scholar] [CrossRef]
  26. Makhoul, J. Linear prediction: A tutorial review. Proc. IEEE 1975, 63, 561–580. [Google Scholar] [CrossRef]
  27. Hayes, M.H. Statistical Digital Signal Processing and Modeling; John Wiley & Sons: Hoboken, NJ, USA, 1996. [Google Scholar]
  28. Fayek, H.M. Speech Processing for Machine Learning: Filter banks, Mel-Frequency Cepstral Coefficients (MFCCs) and What’s In-Between. 2016. Available online: https://haythamfayek.com/2016/04/21/speech-processing-for-machine-learning.html (accessed on 21 April 2016).
  29. Stevens, S.S.; Volkmann, J.; Newman, E.B. A scale for the measurement of the psychological magnitude pitch. J. Acoust. Soc. Am. 1937, 8, 185–190. [Google Scholar] [CrossRef]
  30. Rabiner, L.; Juang, B.H. Fundamentals of Speech Recognition; Prentice-Hall, Inc.: Upper Saddle River, NJ, USA, 1993. [Google Scholar]
  31. ISO/IEC 11172-3; Information Technology—Coding of Moving Pictures and Associated Audio for Digital Storage Media at up to About 1.5 Mbit/s—Part 3: Audio. Standard for MPEG-1 Audio Layer I/II/III (Including MP3); ISO: Geneva, Switzerland, 1993.
  32. Wan, L.; Wang, Q.; Papir, A.; Moreno, I.L. Generalized End-to-End Loss for Speaker Verification. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 4879–4883. [Google Scholar]
  33. Hannun, A.; Case, C.; Jared Casper, B.C.; Diamos, G.; Erich Elsen, R.P.; Satheesh, S.; Sengupta, S.; Coates, A.; Ng, A.Y. Deep Speech: Scaling up end-to-end speech recognition. arXiv 2014, arXiv:1412.5567. [Google Scholar]
  34. Li, L.; Liu, M.; Yao, Y.; Dang, F.; Liu, Y. Patronus: Preventing unauthorized speech recordings with support for selective unscrambling. In Proceedings of the 18th ACM Conference on Embedded Networked Sensor Systems (SenSys’20), Virtual Event, 16–19 November 2020. [Google Scholar]
  35. Hao, X.; Su, X.; Horaud, R.; Li, X. Fullsubnet: A Full-Band and Sub-Band Fusion Model for Real-Time Single-Channel Speech Enhancement. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 6633–6637. [Google Scholar]
Figure 1. Nonlinear transformation in microphone system: (a) The signal path of the voice acquisition device includes a microphone, an amplifier, a low-pass filter, and an ADC. (b) Out-of-band inaudible signals in the voice recording system. (c) Utilizing the nonlinear characteristics of the microphone, out-of-band inaccessible signals can be converted into in-band audible signals.
Figure 1. Nonlinear transformation in microphone system: (a) The signal path of the voice acquisition device includes a microphone, an amplifier, a low-pass filter, and an ADC. (b) Out-of-band inaudible signals in the voice recording system. (c) Utilizing the nonlinear characteristics of the microphone, out-of-band inaccessible signals can be converted into in-band audible signals.
Electronics 14 02960 g001
Figure 2. System threat model.
Figure 2. System threat model.
Electronics 14 02960 g002
Figure 3. System architecture.
Figure 3. System architecture.
Electronics 14 02960 g003
Figure 4. TFM interference system bench test device.
Figure 4. TFM interference system bench test device.
Electronics 14 02960 g004
Figure 5. Simplified schematic of experiment.
Figure 5. Simplified schematic of experiment.
Electronics 14 02960 g005
Figure 6. The time waveform and time–frequency spectrum of raw speech, the speech jammed by noise in Backdoor [17] and the speech jammed by TFM.
Figure 6. The time waveform and time–frequency spectrum of raw speech, the speech jammed by noise in Backdoor [17] and the speech jammed by TFM.
Electronics 14 02960 g006
Figure 7. SNR results of different jamming methods at different distances. The distance refers to the distance from the ultrasonic array to the sound-receiving device.
Figure 7. SNR results of different jamming methods at different distances. The distance refers to the distance from the ultrasonic array to the sound-receiving device.
Electronics 14 02960 g007
Figure 8. Recognition results of several STT systems. (a) WER of Tencent STT. (b) WER of Xunfei STT. (c) WER of Diyun STT. (d) WER of DeepSpeech STT.
Figure 8. Recognition results of several STT systems. (a) WER of Tencent STT. (b) WER of Xunfei STT. (c) WER of Diyun STT. (d) WER of DeepSpeech STT.
Electronics 14 02960 g008
Figure 9. Recognition results after speech enhancement. (a) WER of Tencent STT after speech enhancement. (b) WER of Xunfei after speech enhancement. (c) WER of Diyun STT after speech enhancement. (d) WER of DeepSpeech STT after speech enhancement.
Figure 9. Recognition results after speech enhancement. (a) WER of Tencent STT after speech enhancement. (b) WER of Xunfei after speech enhancement. (c) WER of Diyun STT after speech enhancement. (d) WER of DeepSpeech STT after speech enhancement.
Electronics 14 02960 g009
Figure 10. Recognition results of multiple speakers. (a) The WER of Tencent STT in the case of multiple speakers. (b) The WER of Xunfei STT in the case of multiple speakers. (c) The WER of Diyun STT in the case of multiple speakers. (d) The WER of DeepSpeech STT in the case of multiple speakers.
Figure 10. Recognition results of multiple speakers. (a) The WER of Tencent STT in the case of multiple speakers. (b) The WER of Xunfei STT in the case of multiple speakers. (c) The WER of Diyun STT in the case of multiple speakers. (d) The WER of DeepSpeech STT in the case of multiple speakers.
Electronics 14 02960 g010
Figure 11. Recognition results of different genders.
Figure 11. Recognition results of different genders.
Electronics 14 02960 g011
Table 1. Average recognition results at different distances.
Table 1. Average recognition results at different distances.
Distance (cm)Recognition Results: Avg WER (%)
Noise in Patronus [34]Noise in Backdoor [17]TFM
2062.141.789.3
4035.225.960.3
6015.35.639.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, Z.; Tang, L.; Wang, K.; Tang, X.; Ge, H. Dynamic Ultrasonic Jamming via Time–Frequency Mosaic for Anti-Eavesdropping Systems. Electronics 2025, 14, 2960. https://doi.org/10.3390/electronics14152960

AMA Style

Yu Z, Tang L, Wang K, Tang X, Ge H. Dynamic Ultrasonic Jamming via Time–Frequency Mosaic for Anti-Eavesdropping Systems. Electronics. 2025; 14(15):2960. https://doi.org/10.3390/electronics14152960

Chicago/Turabian Style

Yu, Zichuan, Lu Tang, Kai Wang, Xusheng Tang, and Hongyu Ge. 2025. "Dynamic Ultrasonic Jamming via Time–Frequency Mosaic for Anti-Eavesdropping Systems" Electronics 14, no. 15: 2960. https://doi.org/10.3390/electronics14152960

APA Style

Yu, Z., Tang, L., Wang, K., Tang, X., & Ge, H. (2025). Dynamic Ultrasonic Jamming via Time–Frequency Mosaic for Anti-Eavesdropping Systems. Electronics, 14(15), 2960. https://doi.org/10.3390/electronics14152960

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop