Next Article in Journal
Smart Image Enhancement Using CLAHE Based on an F-Shift Transformation during Decompression
Previous Article in Journal
A Frequency and Radiation Pattern Combo-Reconfigurable Novel Antenna for 5G Applications and Beyond
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reformulating the Binary Masking Approach of Adress as Soft Masking

by
Ruairí de Fréin
School of Electrical and Electronic Engineering, Technological University Dublin, Dublin D08 NF82, Ireland
Electronics 2020, 9(9), 1373; https://doi.org/10.3390/electronics9091373
Submission received: 26 June 2020 / Revised: 14 August 2020 / Accepted: 19 August 2020 / Published: 25 August 2020
(This article belongs to the Section Circuit and Signal Processing)

Abstract

:
Binary masking forms the basis for a number of source separation approaches that have been successfully applied to the problem of de-mixing music sources from a stereo recording. A well-known problem with binary masking is that, when music sources overlap in the time-frequency domain, only one of the overlapping sources can be assigned the energy in a particular time-frequency bin. To overcome this problem, we reformulate the classical pan-pot source separation problem for music sources as a non-negative quadratic program. This reformulation gives rise to an algorithm, called Redress, which extends the popular Adress algorithm. It works by defining an azimuth trajectory for each source based on its spatial position within the stereo field. Redress allows for the allocation of energy in one time-frequency bin to multiple sources. We present results that show that for music recordings Redress improves the SNR, SAR, and SDR in comparison to the Adress algorithm.

1. Introduction

Audio source separation is the problem of extracting sources that have been combined to form an observed audio mixture. Applications include speech enhancement and recognition [1] and hearing aid devices [2]. The separation of music sources from audio mixtures, namely Music Source Separation (MSS), has applications, such as (1) creating audio effects for the purpose of remixing and DJing; (2) automatic transcription of pitched instruments; and, (3) key-signature and chord-detection [3]. Time-Frequency (TF) two-channel methods for audio source separation have left an indelible mark on the field [4,5,6]. These methods are based on one or both of the following estimators, weighted, power-weighted or non-weighted relative attenuation, and/or delay estimation—a unifying framework for these estimators is given in [7]. We consider this mixing case, as most commercial music is in stereo format according to [3] (Formats such as 5.1 are not considered). Stereo mixtures allow for the spatial positioning of sources in a sound field; sources are perceived as originating from the left, centre or right, etc. Positioning is achieved via pan-pot mixing, which scales the contribution of each source to each channel. The relative delay between channels is less important for MSS [5] than speech source separation [4], which uses both relative delay and attenuation estimates to de-mix sources.
Music sources are typically sparse. This means that in the majority of TF bins the source has little energy. The source’s TF support is the remaining TF bins; they are small in number. Sparsity of the underlying sources coupled with independence of occurrence of sources should mean than in general sources do not overlap in TF—the success of Adress [5] bears testament to the conjecture that this is approximately true. However, two-channel TF methods do not generally work well in the TF bins where the sources overlap. Overlap arises for a number of reasons in MSS: harmonic signals are composed of a fundamental frequency and its harmonics; harmony causes a large degree of overlap in the frequency content of sources locally in time; percussive signals have a flatter spectrum than harmonic instruments, which increases the likelihood of them overlapping with harmonic instruments. In this paper, we consider what to do in the TF bins where the music sources do overlap, using a demixing algorithm, called Adress, which has seen wide-spread usage (since it was originally proposed in [5]), and has been the subject of recent analysis in [8]. Before we delve into our contribution, we chart the progress of Adress since it was originally proposed. The Adress algorithm was licensed for use in Sony’s Singstar on the Sony Playstation 3. It was then licensed to Riffstation, which was then sold to Fender and used by millions of users from 2012 to 2018. Our contribution is a soft-masked version of Adress, namely Redress, which seeks to remedy problem of what to do when sources overlap in TF.
The approaches above aim to separate sources by modeling the source positions [4,5,9]. They rely on binary masking and the assumption that only one source has significant energy in a given TF bin. Separation is achieved by assigning each TF bin to one source. This has been shown to be a special case of generalized Wiener filtering [10]. However, if sources overlap, only one source can be assigned the energy in that TF bin. An alternative to these approaches is methods that separate by modeling the sources. Approaches have also leveraged source-filter models in order to over-come the requirement for training of MSS systems. For example, to separate harmonic and percussive sources via tensor factorizations, the authors leveraged source-filter models for pitched instruments as well as constraints to encourage temporal continuity on pitched sources in [11]. Source-filter models were used in order to address the problem of pre-training, which they identified as a shortcoming of the approaches [12,13]. In addition, a disadvantage of pre-training approaches, such as [11], is that they can be processor and memory intensive. More recently, Deep Neural networks (DNNs) have been proposed in order to estimate non-negative masks, which are then applied to the TF representations of the audio mixture [14]. The advantage of TF approaches is that they make it possible to exploit the spectral structure of sources, however, approaches that operate in the time domain directly, such as [15,16], provide a good counter-point.
In contrast to stereo approaches, single-channel recordings do not provide information about the spatial position of sources. Non-negative Matrix Factorization (NMF) [17], which is generally applied to the magnitude spectrogram [18], relies on similar assumptions in terms of the sparsity of the sources in the TF domain to the approaches [4,5,9]. Redress considers the stereo TF instantaneous mixing model proposed in [5]. It does not involve pre-training but instead looks to exploit an initial estimate of the pan-position of the sources to simplify the separation problem by posing it as a non-negative quadratic program. For the squared Euclidean distance objective function, if one of NMF factors is fixed, the resulting problem is a nonnegative quadratic programme. One of the contributions of the present submission is that the Redress nonnegative quadratic programme is constructed by using pre-computed azimuth trajectories as one of the factors. A non-negative matrix factorization approach, where both factors were alternated, was proposed in [19]. This approach was only successful when the number of sources equalled (or was less than) the number of channels; its suitability was limited as there are generally more sources than sensors in practical applications. Redress then solves this optimization problem using a standard NMF-style one-sided update, whose performance is well understood. It works well when there are more sources than sensors.
This paper is organized, as follows. We introduce a simple pan-de-mixing problem to introduce Adress in Section 2. We introduce the main shorting-coming of the Adress approach using this problem. We then use this de-mixing problem to motivate Redress in Section 3. We describe our source reconstruction algorithm in Section 4. Finally, we evaluate the performance of Redress and compare it with Adress in Section 5.

2. Mixing Model

We define pan-mixing using a specific example to establish the notation used in the rest of this paper. Two continuous-time sound sources are considered, namely s 1 ( t ) and s 2 ( t ) . Each source has two frequency components. The first source, s 1 ( t ) , is composed of the frequencies f 1 = 100 Hz and f 3 = 300 Hz and the second source, s 2 ( t ) , is composed of the frequencies f 2 = 200 Hz and f 3 = 300 Hz in Figure 1 (rows 1 and 2). Both sources are scaled by 2 to simplify notation, which yields the expressions
s 1 ( t ) = 2 sin ( 2 π f 1 t ) + 2 sin ( 2 π f 3 t ) ,
s 2 ( t ) = 2 sin ( 2 π f 2 t ) + 2 sin ( 2 s π f 3 t ) .
A stereo mixture of these sources is produced by pan-mixing, by weighting the contribution of each source on the left channel and the right channel. The left and right channel signals, namely x 1 ( t ) and x 2 ( t ) , produced using the weights α and γ , are defined as
x 1 ( t ) = s 1 ( t ) + α s 2 ( t ) ,
x 2 ( t ) = γ s 1 ( t ) + s 2 ( t ) ,
and are illustrated in Figure 1 (rows 3 and 4). The weights lie in the intervals 0 < α < 1 and 0 < γ < 1 .
Appealing to the Fourier transform we can express the source signals, s 1 ( t ) and s 2 ( t ) , more compactly in the frequency domain. The sources are denoted S 1 ( f ) and S 2 ( f ) , where f denotes frequency and where, δ ( · ) , is the delta function,
S 1 ( f ) = δ ( f f 1 ) + δ ( f + f 1 ) + δ ( f f 3 ) + δ ( f + f 3 ) , S 2 ( f ) = δ ( f f 2 ) + δ ( f + f 2 ) + δ ( f f 3 ) + δ ( f + f 3 ) .
As a consequence, the mixtures, X 1 ( f ) and X 2 ( f ) , can also be expressed compactly as,
X 1 ( f ) = δ ( f f 1 ) + δ ( f + f 1 ) + α δ ( f f 2 ) + δ ( f + f 2 ) + ( 1 + α ) δ ( f f 3 ) + δ ( f + f 3 ) ,
X 2 ( f ) = γ δ ( f f 1 ) + δ ( f + f 1 ) + δ ( f f 2 ) + δ ( f + f 2 ) + ( 1 + γ ) δ ( f f 3 ) + δ ( f + f 3 ) ,
in the frequency domain.
These sources have been designed to include one frequency component that causes overlap in the mixtures and a second component that does not overlap in the mixtures.

2.1. Adress

In the Adress algorithm, the authors construct a frequency-azimuth plane as a first step towards separating the sources, s 1 ( t ) and s 2 ( t ) , from the mixtures, x 1 ( t ) and x 2 ( t ) . The frequency-azimuth plane is constructed by varying an independent variable, g, over the range, 0 g 1 and computing the magnitude of the difference between the two frequency domain mixtures. It is necessary to perform this scaling twice, where the roles of X 1 ( f ) and X 2 ( f ) are swapped, to preserve the symmetry of the frequency-azimuth plane. The gwo halves of the frequency-azimuth plane are produced by computing
A 1 ( f ) = | X 1 ( f ) g X 2 ( f ) | ,
A 2 ( f ) = | X 2 ( f ) g X 1 ( f ) | .
Concatenating the components, A 1 ( f ) and A 2 ( f ) , produces the entire frequency-azimuth plane, which is defined as
A ( f ) = A 1 ( f ) A 2 ( f ) .
Given this concatenation of components and the role of g, which can assume values in the range 0 g 1 , in each of these components, when we plot the frequency-azimuth plane we denote the range of g to be 1 g 1 to capture this symmetry. In addition, we refer to the frequency-azimuth plane as the azimugram in the following sections. To de-mix the pan-mixed mixtures presented above, we only need to consider three frequency components f 1 , f 2 and f 3 . The azimugram can be defined in closed form for these three components.
A 1 ( f ) = | 1 g γ | , when f = f 1 , | α g | , when f = f 2 , | ( 1 + α ) g ( 1 + γ ) | , when f = f 3 , 0 , otherwise .
A 2 ( f ) = | γ g | , when f = f 1 , | 1 g α | , when f = f 2 , | ( 1 + γ ) g ( 1 + α ) | , when f = f 3 , 0 , otherwise .
In Equations (11) and (12) the gain satisfies the relationship | g | 1 . The locations of nulls in the azimugram are important for source separation. For the three frequencies f 1 , f 2 and f 3 the null locations are given by the gains
A ( f ) = 0 if g = γ , when f = f 1 , g = α , when f = f 2 , g = 1 + α 1 + γ or 1 + γ 1 + α s . t . g 1 when f = f 3 .
The operation of the Adress algorithm is now summarized for discrete-time source signals which have been sampled at a rate which satisfies the Nyquist–Shannon sampling theorem. Adress computes windowed Discrete Fourier Transforms, X 1 ( k ) and X 2 ( k ) , of the left and right channel mixture signals. The index k denotes the frequency bin indices, k = 0 , 1 , , K . Figure 2 illustrates the positive frequencies of the magnitude spectra of the sources and mixtures, illustrating the frequency components that do overlap and those that do not. We use a set of gain values, g = 0 β , 1 β , β β , where β = M 2 , to construct the frequency-azimuth matrix, A R K × M . The matrix A is generally transformed in order to produce peaks at the locations of nulls in A , yielding the matrix A ^ ,
A ^ [ k , m ] = r [ k ] , if A [ k , m ] min { A [ k , : ] } 0 , otherwise ,
where we introduce r [ k ] = max { A [ k , : ] } min { A [ k , : ] } to simplify the presentation. The motivation for this transformation is to simplify null/peak detection for the user. Figure 3 illustrates the resulting matrix, A ^ , for the two-source mixture introduced above. The peaks located at { f , g } = { 100 , 0.35 } and { 200 , 0.4 } correspond to the frequency components of the sources that do not overlap. Recall that the sign of g is changed in order to distinguish between the right and left components of the azimugram (the value 0.35 is a gain of 0.35 and the negative value indicates which TF mixture is scaled relative to the other one). Reconstruction of the component sources is achieved by assigning TF bins to sources depending on the location of the nulls in the frequency-azimuth plane using an azimuth subspace width parameter H.

2.2. Problem

A null is produced at the gain location, g = α , in the frequency-azimuth plane by the mixture signals described above. This null corresponds to the source, s 2 ( t ) , in the mixture, which excites the frequency f 2 = 200 Hz (cf. Figure 3). The source s 2 ( t ) also excites the frequency f 3 = 300 Hz. The location of the null in the f 3 = 300 Hz frequency band is generally not located at the same gain as the f 2 = 200 Hz component, e.g., α . It is located at either 1 + α 1 + γ or 1 + γ 1 + α depending on the scale of α and γ . In Figure 3, this peak is located at 1 + 0.35 1 + 0.4 0.96 . Note that we have used the scale value 0.35 in the numerator and not 0.35 as the negative sign is used to facilitate plotting both components of the azimugram size-by-side in Figure 3. The scaling of the source signals in the mixtures produce another null at g = γ at the frequency f 1 = 100 Hz. This null arises due to the source, s 1 ( t ) , which is scaled by γ in the second mixture x 2 . A null is generally not located at the gain g = γ in the f 3 = 300 Hz band. Co-occupation of the f 3 bin by the two source signals (which are scaled by different gains) causes the Adress algorithm to assign all of the energy of the frequency f 3 to one of the sources. The other source receives none of the energy. The decision on which source obtains the energy is made based on the distance of the location of the null of the f 3 component from the location of the nulls for the other frequency components of the s 1 and s 2 source signals. For example, in Figure 4, the recovered magnitude spectra for the two sources, obtained using the Adress algorithm are illustrated. The absence of the f 3 frequency component in one of the signals is a deficit. In this example, 50% of the frequency components of one of the sources will be missing and the other source will have a magnitude that is too large for that missing frequency component, because that source has incorrectly been assigned all of the energy in the f 3 frequency bin. This type of problem has been identified and called Frequency-Azimuth Smearing in [5], but it has not been solved.
Methods for trying to ensure that overlap does not happen rely on measuring the level of disjointness of sources in the TF domain; a common measure is Windowed Disjoint Orthogonality (WDO) [4,6]. WDO has been used in order to determine what parametrization of the Short-Time Fourier Transform (STFT) will give the most non-overlapping representation of the source signals in the TF mixtures [4,6]. We now present a solution to the problem of overlapping frequency components presented above that is motivated by re-considering the Adress mixing model.

3. Separation via Azimuth Trajectories

We introduce Redress, which uses azimuth trajectories to recover the source magnitude spectra.
Definition 1.
An azimuth trajectory is the response recorded in the frequency-azimuth matrix, A , as a function of the gain, g, which has been varied over its range 0 g 1 in Equations (8) and (9).
We only consider the positive frequencies due to the symmetry of the TF representation. In the two-source case
X 1 ( f ) = δ ( f f 1 ) + α δ ( f f 2 ) + ( 1 + α ) δ ( f f 3 ) ,
X 2 ( f ) = γ δ ( f f 1 ) + δ ( f f 2 ) + ( 1 + γ ) δ ( f f 3 ) .
In this discussion, one source dominates on the left-channel and the other dominates on the right-channel in order to address both scenarios. The first component of the azimugram is defined as:
A 1 ( f , m ) = X 1 ( f ) g m X 2 ( f ) = ( 1 g m γ ) δ ( f f 1 ) + ( α g m ) δ ( f f 2 ) + [ ( 1 g m γ ) + ( α g m ) ] δ ( f f 3 )
The pan-mixing model does not delay sources and, thus, there is no relative delay between the two microphones. Each term can be expressed as a real-valued scalar, c, times a complex value, z = ( a + b j ) . The absolute value of the product | c z | can be re-expressed as the product of the absolute value of the scalar times the absolute value of the complex value, | c z | = | c | | z | . Two cases warrant examination.
  • When f = f 1 the azimugram can be simplified
    A 1 ( f 1 , m ) = | ( 1 g m γ ) δ ( f 1 f 1 ) | = | ( 1 g m γ ) | | δ ( f 1 f 1 ) | ,
    as the product of s 1 ’s azimuth trajectory h 1 ( g ) = | ( 1 g m γ ) | and s 1 ’s spectral content | δ ( f 1 f 1 ) | at that frequency.
  • When f = f 3 the azimugram can be expressed as
    A 1 ( f 3 , m ) = | ( 1 g m γ ) + ( α g m ) | | δ ( f 3 f 3 ) | ,
    however, from the triangle inequality, it holds that
    | ( 1 g γ ) + ( α g ) | | ( 1 g γ ) | + | ( α g ) | .
    Expressing this portion of the azimugram as the product of source trajectories, h 1 ( g ) = | ( 1 g m γ ) | and h 2 ( g ) = | ( α g m ) | , and their corresponding source spectral content is an approximation.
For mixtures consisting of an arbitrary number of sources, who have, in turn, different magnitudes in each of the TF bins, the error introduced by assuming the inequality is in fact equality in Equation (20) depends on the location of the sources that are occupying the bins, the values of α and γ in Equation (20) and also the magnitude spectra in those frequency bins.
Continuing with discrete TF representations of the signals, and cognisant of the approximation introduced by the triangle inequality, we approximate the azimugram component A 1 ( k , m ) with the following factorization in the two source case. The two matrices, W and H , are defined as
A 1 δ k k 1 0 0 δ k k 2 δ k k 3 δ k k 3 × | ( 1 g 0 γ ) | | ( 1 g 1 γ ) | | ( 1 g M 2 γ ) | | ( α g 0 ) | | ( α g 1 ) | | ( α g M 2 ) |
where the bins k 1 , k 2 , and k 3 correspond to the frequencies f 1 , f 2 , and f 3 , and δ k k i is the Kronecker delta. A similar factorization can be constructed for the other component of the azimugram A 2 ( k , m ) .

4. Reconstruction

In the case of an arbitrary number of sources, R, the azimugram A R + K × M is approximated by the product of the source magnitude TF Spectra in the τ -th time window times the azimuth trajectory of that source. The azimuth trajectories matrix is pre-computed. It is formed by adding a row to the matrix H R + R × M for each source, given estimates of the azimuths of each source, a = { a 1 , a 2 , , a R } . These azimuths are either selected by the user (cf. [5]) or estimated while using a relative attenuation estimator (cf. [7]). If the r-th source dominates on one channel the trajectory is
h r = , | 1 g 2 a r | , | 1 g 1 a r | , 1 , | a r g 1 | , | a r g 2 | , .
If the r-th source dominates on the other channel the trajectory is
h r = , | a r g 2 | , | a r g 1 | , 1 , | 1 g 1 a r | , | 1 g 2 a r | , .
The azimuth trajectories matrix, given the azimuth estimates a, is
H = h 1 h 2 h R T .
The recovery of the magnitude spectra of the source signals is achieved by solving the non-negative quadratic programme
min W | | A W H | | 2 2 , subject to W k , m 0 .
where W R + M × R for the azimugram computed for the τ -th window of the analyzed mixtures X 1 [ k , τ ] and X 2 [ k , τ ] . We use the update proposed by Lee and Seung in [17], e.g., W W A H T W H H T . The resulting W factor is approximately equal to the source magnitude TF spectra:
W | S 1 [ 1 , τ ] | | S 2 [ 1 , τ ] | | S R [ 1 , τ ] | | S 1 [ 2 , τ ] | | S 2 [ 2 , τ ] | | S R [ 2 , τ ] | | S 1 [ K , τ ] | | S 2 [ K , τ ] | | S R [ K , τ ] | .
The reconstruction of the discrete-time r-th source is achieved by forming an estimate of its entire magnitude TF Spectrum, by taking the r-th column from each estimate of W and forming
S ^ = | S r [ 1 , 1 ] | | S r [ 1 , 2 ] | | S r [ 1 , τ ] | | S r [ 2 , 1 ] | | S r [ 2 , 2 ] | | S r [ 2 , τ ] | | S r [ K , 1 ] | | S r [ K , 2 ] | | S r [ K , τ ] | .
We use the mixture phase to reconstruct the discrete-time source. Algorithm 1 summarizes the workflow of the Redress algorithm.
Figure 5 illustrates the time-frequency magnitude spectra of both sources learned using Redress. Both sources are now assigned a frequency component in the f 3 = 300 Hz bin. In comparison Adress is unable to assign energy in the f 3 = 300 Hz bin to both source signals. Figure 6 illustrates the separated azimugrams which result from the Redress approach. In summary, the factorization of the azimugram learned by Redress allows both of the component azimugrams to have peaks at the correct gain for the 300 Hz component.
Algorithm 1: Summary of the Redress Algorithm
Result: Separated sources s 1 , s 2 ,
Compute the STFT of both channels, x 1 and x 2 ;
Compute the frequency azimuth plane by constructing A 1 and A 2 (Equations (8) and (9));
Form the azimuth trajectory matrix, H , using Equations (22)–(24);
Recover the sources by solving the NQP in Equation (25);
Extract the sources from the columns of the source magnitude TF Spectra matrix, W , matrix in Equation (26);
Reconstruct the source magnitude spectra using Equation (27);
Use the mixture phase to reconstruct the discrete-time sources;

5. Evaluation

We generated stereo pan-mixed mixtures using up to four of the original stems, (bass, drums, other instruments, and vocals) provided in the Demixing Secret Dataset (DSD) in order to evaluate the performance of Redress [20]. According to [3], the DSD dataset has gained traction as an evaluation dataset for source separation problems. Evaluation was carried out using an FFT size of 4096 samples, with a hopsize of 2048 samples, and a sampling frequency of 44.1 kHz in order to be consistent with [5]. The analysis window used was a Hamming window. We ran the Redress algorithm for 100 iterations. All of the mixtures were created by down-mixing any stereo stems to mono and then remixing them with two, three of four different azimuth positions depending on the number of sources in the mixture. The source-count parameter, R, of Redress was set to the number of sources in the mixture. For both Adress and Redress, we assumed the azimuths a were known. Both Adress and Redress were implemented in Matlab 2018b (9.5.0.944444 (R2018b), Mathworks, Natick, MA, USA).
The purpose of this evaluation was to test the hypothesis that Redress could improve the source estimates achieved by a binary masking approach when the sources overlapped in the TF domain. We increased the number of source signals present in the mixtures from two to four in order to vary the level of overlap in the TF domain. Four was the maximum number of sources available for each track in the DSD dataset. The baseline method used in our comparison was the Adress algorithm, which uses a binary masking approach for separation.

5.1. Separation Example: Redress

We present a separation example using a four-source mixture and the Redress algorithm. Figure 7 illustrates a 10 s excerpt of the original four source components of Patrick Talbot’s “Set Me Free” from the DSD dataset. Two components are panned left and the other two components are panned right. Figure 8 illustrates the estimates of these sources achieved by Redress. The separations are of high quality; there are very few visual differences between the original sources and the estimates of them. The bass, other instruments, and voice waveforms contain little or no evidence of the percussion/drums events. Similarly, the drums waveform contains little or none of the components of the harmonic instruments.
We illustrate the magnitude TF representations of the original sources and estimates of these sources signals in Figure 9 and Figure 10. The drums have significant spectral energy from 0–15 kHz. The voice and instruments are compactly supported in the range 0–5 kHz. Both the other instruments and the voice have energy in the frequency range 5–15 kHz. This energy overlaps with much of the drums’ energy. Figure 9 illustrates that estimates of the sources: drums, voice, and other instruments, have energy in this range, which illustrates the benefit of Redress over a binary masking approach. The bass and other instruments have been assigned less of this energy. In a binary mask-based source separation approach, only one source can be assigned energy in a TF bin, which can cause the resulting magnitude TF spectra to have gaps if that energy is assigned to another source.
On listening to the recovered waveforms, we conclude that the quality is high. There are some audible artifacts where ideal separation is not achieved, which we now describe. Some traces of the higher frequencies of the drums can be heard on one other estimated source, the other instruments waveform. Some components of the voice waveform are present in the bass waveform. However, the drums waveform is very good; both the low and high frequency components are present. Similarly, the voice waveform is excellent in spite of the fact that some of the voice higher frequency energy is assigned to the bass. This experiment suggests that TF bins can be assigned to multiple source estimates, which is not possible using binary masking approaches. The Redress algorithm seeks to minimize the reconstruction error and, to do this, it may assign the energy in TF to multiple sources. From the triangle inequality (Equation (20)), the factorization that we have proposed is not exact. It is exact when sources are disjoint in TF; it is approximately correct when sources are not disjoint, and the quality of the approximation depends on the magnitude of the sources sharing the TF and their panning weights. We now investigate the approximation achieved by Redress and the binary masking approach used by Adress.

5.2. Overlapping Sources in TF

Table 1, Table 2 and Table 3 illustrate the Signal-to-Noise Ratios (SNRs) of the source estimates achieved over a range of mixing scenarios. In the first case, two of the four sources are mixed and the other two sources are not included in the mixtures (cf. Table 1). In Table 1, the SNRs for estimates of the two sources in the mixtures are indicated and a dash indicates that the corresponding source was not present in the mixture. Similarly, in the second case, three of the four sources are included in the mixture (cf. Table 2) and in the final case all four sources are included Table 3. In general, the SNRs of the estimated sources decreases when the number of sources increases, which can be attributed to the increased TF overlap of the sources when more sources are present. This first result demonstrates that Redress is affected by increased TF overlap. On average, Redress improves the SNR of the reconstructed sources by 3, 1 and 0.1 dB in the two, three, and four sources mixtures in comparison with the sources recovered by Adress. Note that the results for the Adress algorithm are improved by artificial means in these experiments. When the sources are reconstructed by Adress, we have the choice of reconstructing from the left channel or the right channel. The choice of the left channel over the right channel has a significant bearing on the SNR of the result. To compare Redress with the best possible Adress result, we use knowledge of the original sources and estimate the SNR for sources, which have been estimated from both channels and choose the maximum value. In the Redress case, we do not leverage knowledge about the true sources in order to improve the estimate achieved. Even with this advantage, Redress outperforms Adress in terms of the SNR estimates of the recovered sources. In these experiments, the Adress algorithm was parameterized with 201 equally spaced azimuth positions and a target azimuth range of 20 azimuth positions. These parameters were used in the evaluation of Adress in [8].
Figure 11 displays boxplots of the measurements obtained for each mixture class, e.g., 2, 3, or 4 sources while using the MATLAB toolbox BSS EVAL [21]. The measurements considered included (1) the Source-to-Distortion Ratio (SDR) which we interpret as a global quality assessment; (2) the Source-to-Artifacts Ratio (SAR) which in this case is related to the level of musical noise introduced into the source estimates; and finally, (3) the Source-to-Interference Ratio (SIR), which measures the interference from other sources in the estimated sources. In general, all of the measurements decreased as the number of sources increased due the increasing likelihood of sources overlapping in TF. Adress yielded higher Source-to-Interference Ratios (SIR) than Redress. This is unsurprising, as Redress was developed to be able to assign contributions in TF bins to multiple sources. This has the consequence of decreasing SIRs achieved by it when compared to the binary masking approach, Adress. A disadvantage of binary masking approaches is that the aggressive way that TF bins are assigned to one source only typically causes the level of distortion and artefacts that are introduced into the source estimates to increase. Figure 11 illustrates that Redress generally yields higher SDRs and SARs than Adress. The SDR and SAR capture the overall sound quality of the separated source signals. Artifacts—mainly musical noise—are reduced by Redress’s ability to assign TF bins to multiple sources. As a consequence of this, the TF spectra of the sources tend to have fewer isolated bins with energy. This is evident from the TF representations of the separated sources in Figure 9.

5.3. Discussion and Future Work

One of the main contributions of this paper was to reformulate the binary masking approach of Adress as a soft masking problem and then to solve this problem using a nonnegative quadratic programme. The SAR, SNR, and SDR scores achieved by Redress were better than the Adress algorithm under the condition that the number of sources was known. In the experiments, not all sources were playing at the same time and with the same intensity. The challenge of time varying numbers of sources was also faced by the Adress algorithm. It would be interesting to consider how both algorithms could adaptively set the number of sources in order to improve their performance when the assumed number of sources was incorrect. In many cases, the Adress algorithm implementation allows for the user to choose target azimuths for sources as a way to solve this problem. The Redress algorithm could also benefit from similar user supervision.
Regarding the SIR performance of both algorithms, the SIR performance was better for the original Adress algorithm than the proposed algorithm, Redress. This would imply that Adress provides better separation of the sources. Thus, Redress seems to make separation worse, but the overall sound quality better. It is important to consider the target application of Redress. The separated sources learned by Adress are typically re-mixed in order to reduce the effects of the musical noise introduced by the algorithm as a consequence of the binary masking approach it takes to source separation. In effect, Adress separates the sources and then remixes the sources in order to reduce the artifacts and distortions (measured using SARs and SDRs) that arise as a consequence of the uncompromising nature of binary masking.
One of the motivations for the interest in Adress was its low computation complexity and, thus, the possibility of implementing it in real-time. In many cases, Adress is not used in real-time, as the user selects an appropriate azimuth based on an inspection of a simplified form of the azimugram. In comparison, our implementation of Redress in the current paper uses 100 iterations of an NMF-style update, which was proposed by Lee and Seung, e.g., W W A H T W H H T , in order to reconstruct the source signals. A first analysis of the computational complexity of the Redress approach can be summarized as follows. We only consider computations that are different to those computed by Adress. The matrix H can be pre-computed by the system for all possible azimuths. Therefore, it can be stored as a look-up table and so it does not pose a computational burden for a real-time implementation of the approach. Similarly, the matrix product, H H T , can be computed off-line and stored in a look-up table for run-time. The computational cost of Redress involves two matrix products, A H T and W H H T , and one matrix element-wise product and division. The complexity of these terms is set by the size of the FFT, the number of sources to detect and the number of azimuth positions. We posit that this computational load should not be a barrier to implementing Redress in a manner that has the same responsive performance as Adress, even when this update is run in an iteration for 100 steps.

6. Conclusions

In this paper, we investigated whether it was possible to use pre-computed source azimuth trajectories as activation functions in a pan-pot de-mixing problem. We showed that by doing so, the de-mixing problem could be reformulated as a Non-negative Quadratic Program which allowed the Redress source separation algorithm to assign energy to multiple sources who shared a TF bin. The results suggested that Redress decreased the artefacts that were introduced into source estimates. This came at the cost of increasing the level of interference from other sources in those estimates. Redress has a number of advantages that we summarize now. The problem that binary masking had with regard to the f 3 frequency, the shared frequency bin in the motivating example, in Section 2, has been addressed. Redress allocates some of the energy of the mixture at f 3 Hz to both sources. The allocation is generally not the correct solution, but it is preferable to allocating all of the energy to one source, and none to the other sources. With regard to the parameter selection that is required for Adress, we selected the value H, which is an azimuth range used to partition the frequency-azimiuth plane, by testing a number of different settings for the best result. Redress only required that the number of sources R be set for it to out-perform Adress. No other testing to optimize parameters was required. In addition, using Redress, there was no ambiguity about which channel should be used to de-mix the sources. When sources are reconstructed, the Adress algorithm can use either mixture to recover the source signals. In our experiments, we reconstructed the sources using both mixture magnitude TF spectra and then picked the one with the best SNR, SIR, SDR, or SAR, depending on the context. Redress estimated the source magnitude spectra using both mixtures and so no choice is available, or required. Finally, Redress outperformed Adress in terms of the achieved SNR, SDR, and SAR for mixtures consisting of two to four sources.

Funding

This publication has emanated from research conducted with the financial support of Science Foundation Ireland (SFI) under the Grant Number 15/SIRG/3459.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Barker, J.; Watanabe, W.; Vincent, E.; Trmal, J. The fifth chime speech separation and recognition challenge: Dataset, task and baselines. arXiv 2018, arXiv:1803.10609. [Google Scholar]
  2. Wang, D. Time-frequency masking for speech separation and its potential for hearing aid design. Trends Amplif. 2008, 12, 332–353. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Cano, E.; FitzGerald, D.; Liutkus, A.; Plumbley, M.; Stöter, F. Musical Source Separation: An Introduction. IEEE Signal Process. Mag. 2019, 36, 31–40. [Google Scholar] [CrossRef]
  4. Yilmaz, O.; Rickard, S. Blind separation of speech mixtures via time-frequency masking. IEEE Trans. Signal Process. 2004, 52, 1830–1847. [Google Scholar] [CrossRef]
  5. Barry, D.; Lawlor, B.; Coyle, E. Sound Source Separation: Azimuth Discrimination and Resynthesis. In Proceedings of the 7th International Conference on Digital Audio Effects, Naples, Italy, 5–8 October 2004. [Google Scholar]
  6. de Fréin, R.; Rickard, S.T. The Synchronized Short-Time-Fourier-Transform: Properties and Definitions for Multichannel Source Separation. IEEE Trans. Signal Process. 2011, 59, 91–103. [Google Scholar] [CrossRef] [Green Version]
  7. de Fréin, R.; Rickard, S.T. Power-Weighted Divergences for Relative Attenuation and Delay Estimation. IEEE Signal Process. Lett. 2016, 23, 1612–1616. [Google Scholar] [CrossRef] [Green Version]
  8. Barry, D. Real-Time Sound Source Separation for Music Applications. Ph.D. Thesis, Technological University Dublin, Dublin, Ireland, 2019. [Google Scholar] [CrossRef]
  9. FitzGerald, D.; Liutkus, A.; Badeau, R. Projection-Based Demixing of Spatial Audio. IEEE/ACM Trans. Audio Speech Lang. Process. 2016, 24, 1560–1572. [Google Scholar] [CrossRef]
  10. Duong, N.; Vincent, E.; Gribonval, R. Under-Determined Reverberant Audio Source Separation Using a Full-Rank Spatial Covariance Model. IEEE Trans. Audio Speech Lang. Process. 2010, 18, 1830–1840. [Google Scholar] [CrossRef] [Green Version]
  11. Fitzgerald, D.; Coyle, E.; Cranitch, M. Using Tensor Factorisation Models to Separate Drums from Polyphonic Music. In Proceedings of the International Conference on Digital Audio Effects (DAFX09), Como, Italy, 1–4 September 2009; pp. 1–4. [Google Scholar]
  12. Yoshii, K.; Goto, M.; Okuno, H.G. Drum Sound Recognition for Polyphonic Audio Signals by Adaptation and Matching of Spectrogram Templates With Harmonic Structure Suppression. IEEE Trans. Audio Speech Lang. Process. 2007, 15, 333–345. [Google Scholar] [CrossRef] [Green Version]
  13. Gillet, O.; Richard, G. Transcription and Separation of Drum Signals From Polyphonic Music. IEEE Trans. Audio Speech Lang. Process. 2008, 16, 529–540. [Google Scholar] [CrossRef] [Green Version]
  14. Wang, D.; Chen, J. Supervised Speech Separation Based on Deep Learning: An Overview. IEEE/ACM Trans. Audio Speech Lang. Process. 2018, 26, 1702–1726. [Google Scholar] [CrossRef] [PubMed]
  15. Luo, Y.; Mesgarani, N. TaSNet: Time-Domain Audio Separation Network for Real-Time, Single-Channel Speech Separation. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 696–700. [Google Scholar]
  16. Luo, Y.; Mesgarani, N. Conv-TasNet: Surpassing Ideal Time–Frequency Magnitude Masking for Speech Separation. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 27, 1256–1266. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Lee, D.; Seung, H. Algorithms for Non-Negative Matrix Factorization; MIT Press: Cambridge, MA, USA, 2000; pp. 556–562. [Google Scholar]
  18. Virtanen, T. Monaural Sound Source Separation by Nonnegative Matrix Factorization with Temporal Continuity and Sparseness Criteria. IEEE Trans. Audio Speech Lang. Process. 2007, 15, 1066–1074. [Google Scholar] [CrossRef] [Green Version]
  19. de Fréin, R. Remedying Sound Source Separation via Azimuth Discrimination and Re-synthesis. In Proceedings of the Irish Signals and Systems Conference, Donegal, Ireland, 11–12 June 2020. [Google Scholar]
  20. Liutkus, A.; Stöter, F.; Rafii, Z.; Kitamura, D.; Rivet, B.; Ito, N.; Ono, N.; Fontecave, J. The 2016 Signal Separation Evaluation Campaign. In Latent Variable Analysis and Signal Separation; Tichavský, P., Babaie-Zadeh, M., Michel, O., Thirion-Moreau, N., Eds.; Springer: Cham, Switzerland, 2017; pp. 323–332. [Google Scholar]
  21. Vincent, E.; Gribonval, R.; Fevotte, C. Performance measurement in blind audio source separation. IEEE Trans. Audio Speech Lang. Process. 2006, 14, 1462–1469. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Rows 1 and 2 illustrate two sources signals. Source 1, s 1 ( t ) , has a 100 Hz and a 300 Hz frequency component. Source 2, s 1 ( t ) , has a 200 Hz and a 300 Hz frequency component. Rows 3 and 4 illustrate pan-mixed mixtures of the sources, x 1 ( t ) and x 2 ( t ) .
Figure 1. Rows 1 and 2 illustrate two sources signals. Source 1, s 1 ( t ) , has a 100 Hz and a 300 Hz frequency component. Source 2, s 1 ( t ) , has a 200 Hz and a 300 Hz frequency component. Rows 3 and 4 illustrate pan-mixed mixtures of the sources, x 1 ( t ) and x 2 ( t ) .
Electronics 09 01373 g001
Figure 2. The magnitude spectra of the source signals, s 1 and s 2 , and mixtures, x 1 and x 2 , are illustrated for 155 analysis windows and over the frequency range 0 f 1000 Hz.
Figure 2. The magnitude spectra of the source signals, s 1 and s 2 , and mixtures, x 1 and x 2 , are illustrated for 155 analysis windows and over the frequency range 0 f 1000 Hz.
Electronics 09 01373 g002
Figure 3. The frequency-azimuth plane constructed for the mixtures x 1 and x 2 are illustrated for the frequency range 0 f 500 Hz and for gains (or azimuths) 1 g 1 . Peaks at are positioned at the gains 0.35 and 0.4 for the s 1 and s 2 at frequencies 100 Hz and 200 Hz, respectively. The 300 Hz component of s 1 and s 2 does not appear as two separate peaks, but as one peak with a gain that is inconsistent with s 1 and s 2 ’s azimuth.
Figure 3. The frequency-azimuth plane constructed for the mixtures x 1 and x 2 are illustrated for the frequency range 0 f 500 Hz and for gains (or azimuths) 1 g 1 . Peaks at are positioned at the gains 0.35 and 0.4 for the s 1 and s 2 at frequencies 100 Hz and 200 Hz, respectively. The 300 Hz component of s 1 and s 2 does not appear as two separate peaks, but as one peak with a gain that is inconsistent with s 1 and s 2 ’s azimuth.
Electronics 09 01373 g003
Figure 4. The reconstruction achieved by Adress using an azimuth subspace width of H = 65 azimuths is illustrated. We use 100 azimuths for the LHS component and 100 azimuths for the RHS component of the azimugram. In the case that H = 65 , only one de-mixed spectra contains the 300 Hz frequency component. The value H = 65 is chosen so that the peak at the frequency f 3 is assigned to the nearest source. The peak at 0.96 is approximately 61 azimuths away from the peak at 0.35 and so it is assigned to the source s 1 . The peak at 0.4 is 75 azimuths away from the peak at 0.35 and so the f 2 frequency component is not assigned to the s 1 source. Choosing a smaller H might result in the f 3 component not being assigned to either source. Choosing H can be a challenge. Secondly, the magnitudes of the components depend on the channel used to de-mix the sources, which is a disadvantage. The intensity of the f 1 component is different relative to the f 3 component depending on which mixture is used to de-mix the sources. Finally, the f 3 component that is assigned to s 1 includes the energy from the s 2 source as well as the s 1 source.
Figure 4. The reconstruction achieved by Adress using an azimuth subspace width of H = 65 azimuths is illustrated. We use 100 azimuths for the LHS component and 100 azimuths for the RHS component of the azimugram. In the case that H = 65 , only one de-mixed spectra contains the 300 Hz frequency component. The value H = 65 is chosen so that the peak at the frequency f 3 is assigned to the nearest source. The peak at 0.96 is approximately 61 azimuths away from the peak at 0.35 and so it is assigned to the source s 1 . The peak at 0.4 is 75 azimuths away from the peak at 0.35 and so the f 2 frequency component is not assigned to the s 1 source. Choosing a smaller H might result in the f 3 component not being assigned to either source. Choosing H can be a challenge. Secondly, the magnitudes of the components depend on the channel used to de-mix the sources, which is a disadvantage. The intensity of the f 1 component is different relative to the f 3 component depending on which mixture is used to de-mix the sources. Finally, the f 3 component that is assigned to s 1 includes the energy from the s 2 source as well as the s 1 source.
Electronics 09 01373 g004
Figure 5. Reconstruction using Redress: both of the estimates, S 1 ( f ) and S 2 ( f ) , have non-zero magnitudes at 300 Hz.
Figure 5. Reconstruction using Redress: both of the estimates, S 1 ( f ) and S 2 ( f ) , have non-zero magnitudes at 300 Hz.
Electronics 09 01373 g005
Figure 6. Separated Azimugrams using Redress: both of the azimugrams have peaks at the correct gain for the 300 Hz component.
Figure 6. Separated Azimugrams using Redress: both of the azimugrams have peaks at the correct gain for the 300 Hz component.
Electronics 09 01373 g006
Figure 7. Original Sources: bass (row 1), drums (row 2), other instruments (row 3), voice (row 4).
Figure 7. Original Sources: bass (row 1), drums (row 2), other instruments (row 3), voice (row 4).
Electronics 09 01373 g007
Figure 8. Redress Source Estimates: bass (row 1), drums (row 2), other instruments (row 3), voice (row 4).
Figure 8. Redress Source Estimates: bass (row 1), drums (row 2), other instruments (row 3), voice (row 4).
Electronics 09 01373 g008
Figure 9. TF Source Estimates: bass (column 1), drums (column 2), other instruments (column 3), voice (column 4).
Figure 9. TF Source Estimates: bass (column 1), drums (column 2), other instruments (column 3), voice (column 4).
Electronics 09 01373 g009
Figure 10. TF Sources Original: bass (column 1), drums (column 2), other instruments (column 3), voice (column 4).
Figure 10. TF Sources Original: bass (column 1), drums (column 2), other instruments (column 3), voice (column 4).
Electronics 09 01373 g010
Figure 11. SDR, SIR, SAR of Redress, and Adress grouped by number of sources in the mixture.
Figure 11. SDR, SIR, SAR of Redress, and Adress grouped by number of sources in the mixture.
Electronics 09 01373 g011
Table 1. Av. SNR (dB) of the source estimates: 2-source mixtures.
Table 1. Av. SNR (dB) of the source estimates: 2-source mixtures.
BassDrumsOther InstrumentsVoice
Redress11.866714.5693--
Adress10.758511.4251--
Redress-13.378415.7878-
Adress-11.441212.1051-
Redress--10.463413.4975
Adress--9.373910.3778
Redress13.0597--16.4422
Adress10.1297--8.7984
Table 2. Av. SNR (dB) of the source estimates: 3-source mixtures.
Table 2. Av. SNR (dB) of the source estimates: 3-source mixtures.
BassDrumsOther InstrumentsVoice
Redress9.888410.87958.7282-
Adress10.04378.62797.4703-
Redress9.030511.5192-8.8494
Adress9.06949.7596-7.3313
Redress10.7974-10.70528.2186
Adress10.8326-8.57267.7733
Redress-10.17729.69427.6430
Adress-10.15827.87177.0842
Table 3. Av. SNR (dB) of the source estimates: 4-source mixtures.
Table 3. Av. SNR (dB) of the source estimates: 4-source mixtures.
BassDrumsOther InstrumentsVoice
Redress6.08548.72915.87954.1330
Adress5.22247.33865.07026.8377

Share and Cite

MDPI and ACS Style

de Fréin, R. Reformulating the Binary Masking Approach of Adress as Soft Masking. Electronics 2020, 9, 1373. https://doi.org/10.3390/electronics9091373

AMA Style

de Fréin R. Reformulating the Binary Masking Approach of Adress as Soft Masking. Electronics. 2020; 9(9):1373. https://doi.org/10.3390/electronics9091373

Chicago/Turabian Style

de Fréin, Ruairí. 2020. "Reformulating the Binary Masking Approach of Adress as Soft Masking" Electronics 9, no. 9: 1373. https://doi.org/10.3390/electronics9091373

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop