10
IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22, NO. 12, DECEMBER 2014 1849 On Training Targets for Supervised Speech Separation Yuxuan Wang, Arun Narayanan,and DeLiang Wang, Fellow, IEEE Abstract—Formulation of speech separation as a supervised learning problem has shown considerable promise. In its simplest form, a supervised learning algorithm, typically a deep neural network, is trained to learn a mapping from noisy features to a time-frequency representation of the target of interest. Tradition- ally, the ideal binary mask (IBM) is used as the target because of its simplicity and large speech intelligibility gains. The supervised learning framework, however, is not restricted to the use of bi- nary targets. In this study, we evaluate and compare separation results by using different training targets, including the IBM, the target binary mask, the ideal ratio mask (IRM), the short-time Fourier transform spectral magnitude and its corresponding mask (FFT-MASK), and the Gammatone frequency power spectrum. Our results in various test conditions reveal that the two ratio mask targets, the IRM and the FFT-MASK, outperform the other targets in terms of objective intelligibility and quality metrics. In addition, we nd that masking based targets, in general, are signicantly better than spectral envelope based targets. We also present comparisons with recent methods in non-negative matrix factorization and speech enhancement, which show clear performance advantages of supervised speech separation. Index Terms—Deep neural networks, speech separation, super- vised learning, training targets. I. INTRODUCTION S PEECH separation, which is the task of separating speech from a noisy mixture, has major applications, such as ro- bust automatic speech recognition (ASR), hearing aids design, and mobile speech communication. Monaural speech separation (i.e., speech separation from single-microphone recordings) is perhaps most desirable from the application standpoint. Com- pared to multi-microphone solutions, a monaural system is less sensitive to room reverberation and spatial source conguration [38]. On the other hand, monaural separation is a severely un- derdetermined gure-ground separation problem. This study fo- cuses on monaural separation. Manuscript received February 16, 2014; revised May 30, 2014; accepted Au- gust 21, 2014. Date of publication August 28, 2014; date of current version September 13, 2014. This work was supported in part by the Air force Of- ce of Scientic Research (AFOSR) under Grant FA9550-12-1-0130, the Na- tional Institute on Deafness and Other Communication (NIDCD) under Grant R01 DC012048, a Small Business Technology Transfer (STTR) subcontract from Kuzer, and the Ohio Supercomputer Center. The associate editor coor- dinating the review of this manuscript and approving it for publication was Prof. Thomas Fang Zheng. Y. Wang and A. Narayanan are with the Department of Computer Science and Engineering, The Ohio State University, Columbus, OH 43210 USA (e-mail: [email protected]; [email protected]). D. Wang is with the Department of Computer Science and Engineering and the Center for Cognitive and Brain Sciences, The Ohio State University, Columbus, OH 43210 USA (e-mail: [email protected]). Color versions of one or more of the gures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identier 10.1109/TASLP.2014.2352935 Monaural speech separation has been widely studied in the speech and signal processing community for decades. From the signal processing viewpoint, many methods have been pro- posed to estimate the ideal Wiener lter, which is the optimal lter to recover clean speech in the minimum mean squared error (MMSE) sense [21]. A popular alternative to Wiener ltering is statistical model-based methods [13], [21], which infer speech spectral coefcients given noisy observations under prior distribution assumptions for speech and noise. Signal processing based methods usually work reasonably well in relatively high signal-to-noise ratio (SNR) conditions. However, they are generally less effective in low SNR and non-stationary noise conditions [21]. In contrast to signal processing based methods, model-based methods build models of speech and/or noise using premixed signals and show promising results in challenging conditions. For example, techniques in [14], [26] build probabilistic in- teraction models between different sources based on learned priors, and show signicant performance gain in low SNR conditions. Another line of work is non-negative matrix fac- torization (NMF) [22], [31], where noisy observations are modeled as weighted sums of non-negative source bases. These model-based methods work well if underlying assumptions are met. However, in our experience (see e.g. [37]), these methods do not generalize well to unseen noisy conditions and are mostly effective for structured interference, e.g. music or a competing speaker. Moreover, these methods often require expensive inference, making them hard to use in real-world speech applications. Recently, we have formulated monaural speech separation as a supervised learning problem, which is a data driven approach. In the simplest form, acoustic features are extracted from noisy mixtures to train a supervised learning algorithm, e.g. a deep neural network (DNN) [36]. In many previous studies (e.g. [10], [16], [17]), the training target (or the learning signal) is set to the ideal binary mask (IBM), which is a binary mask constructed from premixed speech and noise signals (see Section III-A for denition). This simplies speech separation to a binary clas- sication problem, a well studied machine learning task. Fur- thermore, IBM processing has been shown to yield large speech intelligibility improvements even in extremely low SNR condi- tions [1], [2], [19], [33]. Supervised speech separation aiming to estimate the IBM has shown a lot of promise. Notably, this ap- proach has provided the rst demonstration of improved speech intelligibility in noise for both normal hearing [17] and hearing impaired listeners [12]. Supervised speech separation has also been shown to generalize well given sufcient training data [36], [39]. In addition, the system operates in a frame-by-frame fashion and inference is fast, making it amenable to real-time implementation. 2329-9290 © 2014 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND ...web.cse.ohio-state.edu/~wang.77/papers/WNW.taslp14.pdf1850 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22,

  • Upload
    others

  • View
    8

  • Download
    0

Embed Size (px)

Citation preview

Page 1: IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND ...web.cse.ohio-state.edu/~wang.77/papers/WNW.taslp14.pdf1850 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22,

IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22, NO. 12, DECEMBER 2014 1849

On Training Targets for Supervised Speech SeparationYuxuan Wang, Arun Narayanan, and DeLiang Wang, Fellow, IEEE

Abstract—Formulation of speech separation as a supervisedlearning problem has shown considerable promise. In its simplestform, a supervised learning algorithm, typically a deep neuralnetwork, is trained to learn a mapping from noisy features to atime-frequency representation of the target of interest. Tradition-ally, the ideal binary mask (IBM) is used as the target because ofits simplicity and large speech intelligibility gains. The supervisedlearning framework, however, is not restricted to the use of bi-nary targets. In this study, we evaluate and compare separationresults by using different training targets, including the IBM, thetarget binary mask, the ideal ratio mask (IRM), the short-timeFourier transform spectral magnitude and its corresponding mask(FFT-MASK), and the Gammatone frequency power spectrum.Our results in various test conditions reveal that the two ratiomask targets, the IRM and the FFT-MASK, outperform the othertargets in terms of objective intelligibility and quality metrics.In addition, we find that masking based targets, in general, aresignificantly better than spectral envelope based targets. Wealso present comparisons with recent methods in non-negativematrix factorization and speech enhancement, which show clearperformance advantages of supervised speech separation.

Index Terms—Deep neural networks, speech separation, super-vised learning, training targets.

I. INTRODUCTION

S PEECH separation, which is the task of separating speechfrom a noisy mixture, has major applications, such as ro-

bust automatic speech recognition (ASR), hearing aids design,andmobile speech communication.Monaural speech separation(i.e., speech separation from single-microphone recordings) isperhaps most desirable from the application standpoint. Com-pared to multi-microphone solutions, a monaural system is lesssensitive to room reverberation and spatial source configuration[38]. On the other hand, monaural separation is a severely un-derdetermined figure-ground separation problem. This study fo-cuses on monaural separation.

Manuscript received February 16, 2014; revised May 30, 2014; accepted Au-gust 21, 2014. Date of publication August 28, 2014; date of current versionSeptember 13, 2014. This work was supported in part by the Air force Of-fice of Scientific Research (AFOSR) under Grant FA9550-12-1-0130, the Na-tional Institute on Deafness and Other Communication (NIDCD) under GrantR01 DC012048, a Small Business Technology Transfer (STTR) subcontractfrom Kuzer, and the Ohio Supercomputer Center. The associate editor coor-dinating the review of this manuscript and approving it for publication wasProf. Thomas Fang Zheng.Y.Wang and A. Narayanan are with the Department of Computer Science and

Engineering, The Ohio State University, Columbus, OH 43210 USA (e-mail:[email protected]; [email protected]).D. Wang is with the Department of Computer Science and Engineering

and the Center for Cognitive and Brain Sciences, The Ohio State University,Columbus, OH 43210 USA (e-mail: [email protected]).Color versions of one or more of the figures in this paper are available online

at http://ieeexplore.ieee.org.Digital Object Identifier 10.1109/TASLP.2014.2352935

Monaural speech separation has been widely studied in thespeech and signal processing community for decades. Fromthe signal processing viewpoint, many methods have been pro-posed to estimate the ideal Wiener filter, which is the optimalfilter to recover clean speech in the minimum mean squarederror (MMSE) sense [21]. A popular alternative to Wienerfiltering is statistical model-based methods [13], [21], whichinfer speech spectral coefficients given noisy observationsunder prior distribution assumptions for speech and noise.Signal processing based methods usually work reasonablywell in relatively high signal-to-noise ratio (SNR) conditions.However, they are generally less effective in low SNR andnon-stationary noise conditions [21].In contrast to signal processing based methods, model-based

methods build models of speech and/or noise using premixedsignals and show promising results in challenging conditions.For example, techniques in [14], [26] build probabilistic in-teraction models between different sources based on learnedpriors, and show significant performance gain in low SNRconditions. Another line of work is non-negative matrix fac-torization (NMF) [22], [31], where noisy observations aremodeled as weighted sums of non-negative source bases. Thesemodel-based methods work well if underlying assumptionsare met. However, in our experience (see e.g. [37]), thesemethods do not generalize well to unseen noisy conditions andare mostly effective for structured interference, e.g. music ora competing speaker. Moreover, these methods often requireexpensive inference, making them hard to use in real-worldspeech applications.Recently, we have formulated monaural speech separation as

a supervised learning problem, which is a data driven approach.In the simplest form, acoustic features are extracted from noisymixtures to train a supervised learning algorithm, e.g. a deepneural network (DNN) [36]. In many previous studies (e.g. [10],[16], [17]), the training target (or the learning signal) is set to theideal binary mask (IBM), which is a binary mask constructedfrom premixed speech and noise signals (see Section III-A fordefinition). This simplifies speech separation to a binary clas-sification problem, a well studied machine learning task. Fur-thermore, IBM processing has been shown to yield large speechintelligibility improvements even in extremely low SNR condi-tions [1], [2], [19], [33]. Supervised speech separation aiming toestimate the IBM has shown a lot of promise. Notably, this ap-proach has provided the first demonstration of improved speechintelligibility in noise for both normal hearing [17] and hearingimpaired listeners [12]. Supervised speech separation has alsobeen shown to generalize well given sufficient training data[36], [39]. In addition, the system operates in a frame-by-framefashion and inference is fast, making it amenable to real-timeimplementation.

2329-9290 © 2014 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

Page 2: IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND ...web.cse.ohio-state.edu/~wang.77/papers/WNW.taslp14.pdf1850 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22,

1850 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22, NO. 12, DECEMBER 2014

A suitable training target is important for supervised learning.On the one hand, one should use a target that can substantiallyimprove speech perception in noise. On the other hand, the map-ping from features to the target of interest should be amenablefor training, say in terms of optimization difficulty [9]. Althoughthe IBM is the optimal binarymask, it may not necessarily be thebest target for training and prediction. Separation using binarymasking typically produces residual musical noise. Other idealtargets are possible and can potentially improve speech intelli-gibility and/or quality, such as the target binary mask (TBM)[1], [18], the ideal ratio mask (IRM), the short-time Fouriertransform (STFT) spectral magnitude, and the Gammatone fre-quency power spectrum. We note that some of them have beenused in our preliminary work [11], [23], [37]. However, whattraining targets are appropriate for supervised speech separa-tion remains unclear. This is clearly an important question withpotentially important implications for separation performance.This paper addresses this question systematically, including astudy of new training targets. In addition, we compare super-vised separation with NMF and speech enhancement methods.The rest of the paper is organized as follows.We first describe

the DNN based supervised speech separation framework and thevarious training targets that we evaluate in the next two sections.Experimental settings and evaluation and comparison results arepresented in Section IV and V, respectively. Discussions andconclusions are provided in Section VI.

II. SUPERVISED SPEECH SEPARATION

Speech separation can be interpreted as the process that mapsa noisy signal to a separated signal with improved intelligibilityand/or perceptual quality1. Without considering the impact ofphase, this is often treated as the estimation of clean speechmagnitude or some ideal mask. Supervised speech separationformulates this as a supervised learning problem such that themapping is explicitly learned from data. Acoustic features areextracted from a mixture, which, along with the correspondingdesired outputs are fed into a learningmachine for training. Newnoisy mixtures are separated by passing estimated outputs andmixture phase into a resynthesizer.To focus our study on learning targets, we use a fixed set of

complementary features [34] throughout the experiments. Thefeature set includes amplitude modulation spectrogram (AMS),relative spectral transformed perceptual linear prediction co-efficients (RASTA-PLP), mel-frequency cepstral coefficients(MFCC), and 64-channel Gammatone filterbank power spectra(GF). All these features are extracted at the frame level and areconcatenated with the corresponding delta features. We alsoemploy an auto-regressive moving average (ARMA) filter [3],[4] to smooth temporal trajectories of all features:

(1)

Here is the feature vector at frame , is the filteredfeature vector, and is the order of the filter. We use a secondorder ARMA filter ( ), which we found consistently im-proves separation performance in low SNR conditions [4].

1Depending on the application, the desired output of the mapping does nothave to be clean speech.

We use DNNs (multilayer perceptrons) as the discrimina-tive learning machine, which has been shown to work well forspeech separation [35], [36]. All DNNs use three hidden layers,each having 1024 rectified linear hidden units (ReLU) [8]. Thestandard backpropagation algorithm coupled with dropout reg-ularization [15] (dropout rate 0.2) are used to train the networks.No unsupervised pretraining is used. We use the adaptive gra-dient descent [5] along with a momentum term as the optimiza-tion technique. A momentum rate of 0.5 is used for the first 5epochs, after which the rate increases to and is kept as 0.9. TheDNNs are trained to predict the desired outputs across all fre-quency bands, and the mean squared error (MSE) is used as thecost (loss) function. The dimensionality of the output layer de-pends on the target of interest, which is described in the nextsection. For targets in the range [0,1], we use sigmoid activationfunctions in the output layer; for the rest we use linear activationfunctions.To further incorporate temporal context, we splice a 5-frame

window of features as input to the DNNs. The output of the net-work is composed of the corresponding 5-frame window of tar-gets. In other words, the DNNs predict the neighboring frames’targets together. The multiple estimates for each frame are thenaveraged to produce the final estimate. Doing so yields small butconsistent improvements over predicting single-frame targets.

III. TRAINING TARGETS

We introduce six training targets evaluated in this studybelow. We assume that the input signal is sampled at 16 kHz,and use a 20-ms analysis window with 10-ms overlap. Anillustration of different training targets is shown in Fig. 1.

A. Ideal Binary Mask (IBM)

The ideal binary mask is a main computational goal for com-putational auditory scene analysis (CASA) [32]. The IBM is atime-frequency (T-F) mask constructed from premixed signals.For each T-F unit, we set the corresponding mask value to 1 ifthe local SNR is greater than a local criterion (denoted as ),otherwise it is set to 0. Quantitatively, the IBM is defined as:

(2)

where denotes the local SNR within the T-F unitat time and frequency . As mentioned earlier, it is well es-tablished that IBM processing of sound mixtures yields largespeech intelligibility gains for both normal hearing and hearingimpaired listeners. In addition, the effectiveness of IBM estima-tion has been exemplified by the recent success in improvinghuman speech intelligibility [17], [12].Following common practice, we use a 64-channel Gamma-

tone filterbank to derive the IBM; hence the DNN hassigmoidal output units with a 5-frame window of targets.

Note that although we train on binary targets, in testing weuse the posterior probabilities from the DNN, representing theprobability of speech dominance, as a soft mask for resynthesiswhich is found to produce better quality. The choice of hasa significant impact on speech intelligibility [18]; we set tobe 5 dB smaller than the SNR of the mixture to preserve enough

Page 3: IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND ...web.cse.ohio-state.edu/~wang.77/papers/WNW.taslp14.pdf1850 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22,

WANG et al.: ON TRAINING TARGETS FOR SUPERVISED SPEECH SEPARATION 1851

Fig. 1. Various training targets for a TIMIT utterance mixed with a factory noise at dB SNR.

speech information. For example, if the mixture SNR is -5 dB,the corresponding is set to dB.

B. Target Binary Mask (TBM)

Unlike the IBM, the TBM [18] is a binary mask that isobtained by comparing the target speech energy in each T-Funit with a reference speech-shaped noise (SSN). That is, the

term in Eq. (2) is calculated using the referenceSSN, regardless of the actual interference. Although the TBM isobtained in a noise-independent way, subject tests have shownthat the TBM achieves similar intelligibility improvements asthe IBM [18]. The reason the TBM works is that it preservesthe spectrotemporal modulation patterns essential to speechperception, i.e. where the target speech energy occurs in thetime-frequency domain. Since the TBM can be interpreted as acartoon of target speech patterns, it may be easier to learn theTBM than the IBM. We use the same frontend to generate theTBM, i.e. a 64-channel Gammatone filterbank, and the same

values.

C. Ideal Ratio Mask (IRM)

The ideal ratio mask is defined as follows:

(3)

where and denote the speech and noise en-ergy, respectively, in a particular T-F unit. is a tunable pa-rameter to scale the mask. Although technically different, one

can see that the IRM is closely related to the frequency-do-main Wiener filter assuming speech and noise are uncorrelated[28], [21]. We experimented with different values and found

to be the best choice. Interestingly, with ,Eq. (3) becomes similar to the square-root Wiener filter, whichis the optimal estimator of the power spectrum [21].Like the IBM and TBM, the IRM is also obtained by using a

64-channel Gammatone filterbank and is in the range of [0,1].

D. Gammatone Frequency Power Spectrum (GF-POW)

We also evaluate performance by directly predicting the64-channel Gammatone frequency power spectrum (GF-POW)of clean speech. The Gammatone filterbank has a finer reso-lution in lower frequency regions compared to STFT. Sincethere is no direct inverse transformation for Gammatone fil-tering, we convert the estimated power spectrum to a mask,

, for resynthesis. Here, anddenote the speech and mixture energy in the Gammatonefrequency domain, respectively. Predicting GF-POW has beenshown to be useful in our recent supervised dereverberationwork [11]. Note that the construction of the IBM and the IRMalso involves GF-POW.

E. Short-Time Fourier Transform Spectral Magnitude(FFT-MAG) and Mask (FFT-MASK)

If the goal is to recover clean speech, then predicting theSTFT magnitude of clean speech seems to be a very naturalchoice. We use a 320-point FFT analysis; thus the spectral mag-nitude in each frame is a 161- vector. We call this target FFT-MAG. The estimated magnitude combined with the mixturephase is passed through an inverse FFT to generate the estimated

Page 4: IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND ...web.cse.ohio-state.edu/~wang.77/papers/WNW.taslp14.pdf1850 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22,

1852 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22, NO. 12, DECEMBER 2014

clean speech. Raw spectral magnitudes usually have a very largedynamic range, hence proper compression or normalization areneeded to make them amenable to the backpropagation training.We will show in Section V-A that different normalizations canlead to significantly different results. We note that a very recentstudy [39] predicts the log compressed spectral magnitude undera supervised speech separation framework.For a comparison with masking, we straightforwardly rewrite

the clean spectral magnitude, , as

(4)

where is the (uncompressed) noisy STFT spectral mag-nitude. Apart from directly predicting , we also predict

, which can be interpreted as a mask.The clean magnitude is reconstructed by multiplying the pre-dicted mask with the noisy magnitude. We call this target FFT-MASK, which is an intermediate target to recover clean magni-tude. Note that unlike the IRM, the FFT-MASK is not upper-bounded by 1, and therefore we use linear output activationfunctions in the DNNs. For better numerical stability in back-propagation training, we clip values greater than 10 in the FFT-MASK to 10, where 10 is an arbitrarily chosen value. The moti-vation of introducing FFT-MASK is to enable a direct compar-ison with FFT-MAG, as the perfect estimation of these two tar-gets produces the same underlying objective–the clean speechmagnitude.

IV. EXPERIMENTAL SETTINGS

We use 2000 randomly chosen utterances from the TIMIT[7] training set as our training utterances. The TIMIT core testset, which consists of 192 utterances from unseen speakers ofboth genders, is used as the test set. We use SSN and 4 othernoises from the NOISEX dataset [30] as our training and testnoises. These include a babble noise, a factory noise (called“factory1”), a destroyer engine room noise, and an operationroom noise (called “oproom”). Except SSN, all other noises arenon-stationary. All noises are around 4 minutes long. To createthe training sets, we use random cuts from the first 2 minutesof each noise to mix with the training utterances at -5 and 0 dBSNR. The test mixtures are constructed by mixing random cutsfrom the last 2 minutes of each noise with the test utterances at-5, 0 and 5 dB SNR, where 5 dB is an unseen SNR condition. Di-viding the noises into two halves ensures that the new noise seg-ments are used during testing. Aside from the aforementionednoises, we also use an unseen factory noise (called “factory2”)and a tank noise from NOISEX to evaluate generalization per-formance. Note that separation of these broadband noises at lowSNRs is a very challenging task. For example, even for the sta-tionary SSN, the human intelligibility score at -5 dB is onlyaround 65% and 35% for normal hearing and hearing impairedlisteners, respectively [12].To put our results in perspective, we compare with

model-based speech enhancement and NMF based separa-tion. For speech enhancement, we compare with a recentsystem by Hendriks et al. [13], which uses an MMSE estimatorof speech DFT coefficients assuming a generalized gamma

TABLE IPERFORMANCE ON FACTORY1 WHEN THE CLEAN MAGNITUDES ARE

NORMALIZED/COMPRESSED IN DIFFERENT WAYS

distribution for speech magnitude [6]. For noise tracking, astate-of-the-art MMSE noise power spectral density estimatoris used [13]. For a fair comparison with NMF, we use the su-pervised NMF method where the speech bases and noise basesare trained separately (for each type of noise) using the sametraining data used by the DNNs. We made an effort to yield thebest performance of supervised NMF; we have tried differentvariants and found that a recent version using an active-setNewton algorithm (ASNA) [31] produces the best results. Thefinal system, denoted as ASNA-NMF, models a sliding windowof 5 frames of magnitude spectra and uses 160 speech basesand 80 noise bases. Using larger numbers of bases (e.g., 1000)does not seem to improve the performance significantly on ourtest set.For evaluation metrics, we use the Short-Time Objective In-

telligibility score (STOI) [29] to measure the objective intel-ligibility. STOI denotes a correlation of short-time temporalenvelopes between clean and separated speech, and has beenshown to be highly correlated to human speech intelligibilityscore. We also evaluate objective speech quality using the Per-ceptual Evaluation of Speech Quality (PESQ) score [27]. LikeSTOI, PESQ is obtained by comparing the separated speechwith the corresponding clean speech. The STOI score rangesfrom 0 to 1, and PESQ score to 4.5.To supplement the above perceptually oriented metrics, we

also give SNR results, which take into account the underlyingsignal energy of each T-F unit. We should point out that thetraditional SNR metric comparing the separated speech withclean speech is not appropriate here. First, different targets aimto reconstruct different underlying signals. For example, theground truth signal of IBM prediction differs from that of FFT-MASK prediction, therefore the use of the traditional SNR isproblematic. Second, the traditional SNR does not take accountof perceptual effects and it is well documented that SNR maynot correlate with speech intelligibility. For example,dB maximizes the SNR gain of the IBM [20]. However, the

choice of dB is clearly worse than negative LC values(e.g. dB) for both human speech intelligibility [18] and au-tomatic speech recognition performance [24]. In other words,lower output SNRs lead to higher speech intelligibility (see also[17], [12]). As our study focuses on different training targets, itmakes sense to use the target-based SNR that compares the sep-arated speech with the target signal resynthesized from the cor-responding ideal target. That is, the output SNRs of IBM, TBMand IRM predictions are obtained using the signals resynthe-sized from the IBM, TBM and IRM, respectively, as the groundtruth (see also [32]). For FFT-MAG, FFT-MASK, ASNA-NMFand Hendriks et al.’s system, the ground truth signal is resyn-thesized using the clean speech magnitude combined with themixture phase, as the computational objective of these targets/

Page 5: IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND ...web.cse.ohio-state.edu/~wang.77/papers/WNW.taslp14.pdf1850 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22,

WANG et al.: ON TRAINING TARGETS FOR SUPERVISED SPEECH SEPARATION 1853

TABLE IIPERFORMANCE COMPARISONS BETWEEN VARIOUS TARGETS AND SYSTEMS ON dB MIXTURES. “MC-IRM”

STANDS FOR MULTI-CONDITION TRAINING (ON ALL FIVE NOISES) AND USES IRM AS THE TRAINING TARGET

methods is to obtain STFT clean speech magnitude and the sep-arated speech is reconstructed using the mixture phase. Usingthe target-based SNR facilitates comparisons between these twogroups of targets/methods.

V. RESULTS

A. Comparison Between Targets

Before presenting comprehensive evaluations, we comparevarious compression/normalization techniques for predictingFFT-MAG. If one wants to predict the clean magnitude,proper normalizations or compressions are needed becausethe magnitudes typically have a very broad dynamic range,causing difficulty for gradient descent based training algo-rithms. We show that different ways of normalization impactperformance significantly. In Table I, we compare STOI andPESQ performance on the factory 1 noise using different kindsof normalization/compression methods. We first use the logcompression, which is perhaps the most widely used compres-sion technique (e.g. [39]). Since log magnitude is not bounded,we use linear output units in the DNNs. Next, we use percentnormalization, which linearly scales the data to the range of[0,1]. This is done by first subtracting the minimum value,and then dividing by the difference between the maximumand minimum value. We use sigmoidal output units in thiscase. Finally, we normalize the magnitudes by first performinga log compression followed by percent normalization, anduse sigmoidal output units. From Table I we can see that theperformance of these normalization methods does not differtoo much at dB. However, at 0 and 5 dB, the traditionallog compression performs significantly worse in terms of STOI(e.g., 4% worse at 5 dB) than the log compression followedby percent normalization. Using only percent normalizationgives closer, but still worse, STOI results; but its PESQ resultsare the worst among the three. We believe log + percent nor-malization performs better because it preserves spectral detailswhile simultaneously making the target bounded. Therefore,we use this normalization scheme when predicting spectralmagnitude/energy based targets in the remaining experiments.The comparisons between different targets in various test

conditions are shown in Tables II, III, and IV at different mix-ture SNRs, where best score is highlighted by boldface. Inthese tables, “SNR” denotes the target-based SNR mentioned inSection IV. We first discuss the results in the most challengingscenario, the dB SNR case, as shown in Table II. Generally,

regardless of the target of choice, the supervised speech separa-tion framework provides substantial improvements comparedto unprocessed mixtures. For the two binary masking targets,the IBM and the TBM, large improvements are obtained inSTOI and SNR. Although we use the posterior probabilitiesfrom the DNNs as soft masks for resynthesis, the PESQ im-provements are still limited over unprocessed mixtures, exceptin the case of the operation room noise. This is consistent withthe common point of view that binary masking tends to im-prove speech intelligibility but not speech quality. Comparedto the IBM, using the TBM as the target results in similar STOIscores but significantly worse PESQ scores and SNRs. For su-pervised techniques like the one used here, the IBM seems tobe a better choice than the TBM, probably because the TBMis defined by completely ignoring the noise characteristics inthe mixture.Going from binary masking to ratio masking improves

all objective metrics, as exemplified by the performance ofthe IRM. On factory1, babble and SSN, predicting the IRMachieves slightly better or equal STOI results than predictingthe IBM. On the engine and operation room noises, predictingthe IRM yields more than two percent STOI improvements.Predicting the IRM seems to be especially beneficial for im-proving objective speech quality. For example, the PESQ scoreimproves by 0.65 and 0.76 in the engine noise compared tothe IBM and unprocessed mixtures, respectively. On average,predicting the IRM provides a 2 dB SNR improvements overpredicting the IBM.On the two challenging noises, factory1 and babble, and the

relatively easier noise, SSN, predicting FFT-MAG achievessimilar STOI and better PESQ results compared to predictingthe IBM. However, predicting FFT-MAG achieves the worstperformance on the other noises in terms of STOI. For ex-ample, on the operation room noise, the STOI is 3% worsethan predicting the IBM. Similarly, FFT-MAG is consistentlyworse than the IRM, especially in the case of engine noise andoperation room noise.Interestingly, FFT-MASK produces comparable and some-

times even slightly better STOI and PESQ results than the IRM.Also, FFT-MASK produces significantly better SNR resultsthan FFT-MAG on all noises. This contrast with FFT-MAGappears surprising at first, considering that the DNNs in bothcases are essentially trained to estimate the same underlyingtarget, the clean magnitude. We will provide some analysis inthe next subsection as to why FFT-MASK performs better.

Page 6: IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND ...web.cse.ohio-state.edu/~wang.77/papers/WNW.taslp14.pdf1850 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22,

1854 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22, NO. 12, DECEMBER 2014

TABLE IIIPERFORMANCE COMPARISONS BETWEEN VARIOUS TARGETS AND SYSTEMS ON 0 DB MIXTURES

TABLE IVPERFORMANCE COMPARISONS BETWEEN VARIOUS TARGETS AND SYSTEMS ON 5 DB MIXTURES

Predicting GF-POW, which is also a spectral envelopebased target, has a similar performance trend as FFT-MAG. Ingeneral, it produces worse STOI results than either binary orratio masking. Nevertheless, GF-POW seems to be consistentlybetter than FFT-MAG.The performance trend at 0 dB is similar to that at dB, as

shown in Table III. That is to say, all targets improve objectivemetricsoverunprocessedmixtures.Binarymasking significantlyimproves objective intelligibility scores but the improvementin objective quality is minor. Predicting the IRM instead of theIBM significantly improves both objective quality and intelligi-bility metrics. FFT-MAG fails to compete with the other targets,whereas FFT-MASK is on par with the IRM.One noticeable dif-ference at 0 dB is that the performance degradation of FFT-MAGbecomes noticeably larger. For example, in the dB factory1noise condition, FFT-MAG produces the same STOI results asthe IBM, whereas at 0 dB FFT-MAG is 3% points worse.Separation at 5 dB is relatively easier, hence we can see in

Table IV that the STOI difference between various maskingbased targets becomes smaller. In contrast, FFT-MAG performsmuch worse than all the masking based targets. For example, theSTOI and PESQ results obtained on SSN are even worse thanthose of unprocessed mixtures. In general, the IRM and FFT-MASK perform comparably; with the former slightly better onaverage.In Fig. 2, we illustrate the STFT magnitudes of a separated

speech utterance resynthesized using the estimated IBM, IRM,FFT-MAG and FFT-MASK. The mixture here is a TIMIT maleutterance with factory1 at 5 dB. We can see that predictingthe IBM preserves spectrotemporal modulation patterns of theclean speech, which are essential for speech intelligibility [25].Also, the separated speech tends to have clearer onsets/offsetsand sharper spectral transitions. Predicting FFT-MAG worksreasonably well for low frequency regions where most of thespeech energy resides. However, it misses a lot of details in the

mid- to high-frequency regions, which are important for bothintelligibility and quality. Visually speaking, the results are sim-ilar between IRM prediction and FFT-MASK prediction in thesense that they both preserve important modulation patterns aswell as fine structures.

B. Issues with FFT-MAG Prediction

We are interested in why predicting FFT-MAG producesthe worst results, and in particular, why there is a substantialperformance gap between FFT-MAG and FFT-MASK. Westart our analysis with two hypotheses. First, since FFT-MAGis the same across different noises and SNRs, the DNN has tolearn a many-to-one mapping (recall that we train on and0 dB mixtures), which may be a more difficult task comparedto learning a one-to-one mapping as in FFT-MASK. Second,masking may be inherently better. To verify these hypotheses,we designed two experiments. In the first experiment, wetrain a DNN to predict FFT-MAG only for 0 dB factory1 toreduce many-to-one mapping, and in the second experimentwe train a DNN to predict (LOGMASK), whereand denote the clean and noisy magnitude (with time

and frequency index omitted), for and 0 dB factory1.The STOI results are shown in Fig. 3. From the figure wecan see that learning a many-to-one mapping does not seemto be the cause as the performance in the matched SNRcondition is only marginally better. Interestingly, the new mask

does improve performance over FFT-MAG, butis still significantly worse than FFT-MASK. This seems toindicate that, although masking is helpful, the use of logcompression is likely one of the causes of the performancedifference. We further analyze why log compression affectsperformance below.Let denote the ratio between the network’s output and

the desired output. Here , and when or

Page 7: IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND ...web.cse.ohio-state.edu/~wang.77/papers/WNW.taslp14.pdf1850 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22,

WANG et al.: ON TRAINING TARGETS FOR SUPERVISED SPEECH SEPARATION 1855

Fig. 2. STFT magnitudes of a separated speech using different training targets. The mixture here is a TIMIT male utterance mixed with the factory1 noise at 5 dB.

Fig. 3. STOI results for 0 dB factory1. “Matched SNR” denotes trainingFFT-MAG also for 0 dB factory1. “LOGMASK” is a newmask ( ).

, the neural network overestimates or underestimatesthe target, respectively. We assume that the network learns

Fig. 4. Average values across frequency obtained on the training set.

equally well for both targets, meaning that is in the samerange. In Fig. 4, we plot the average values obtained on the

Page 8: IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND ...web.cse.ohio-state.edu/~wang.77/papers/WNW.taslp14.pdf1850 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22,

1856 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22, NO. 12, DECEMBER 2014

Fig. 5. Comparison of and . Left: contour plot of the ratio . Right: normalized error curves ( and ) when.

training set across frequency for both targets, and we can seethat this is basically the case. Since the goal is to estimatethe clean magnitude, we evaluate the estimation error in termsof the absolute deviation from the clean magnitude . Recallthat for FFT-MASK, we predict a mask , whereas for logmagnitude we predict . Therefore for FFT-MASK, thenetwork output is , and the estimation error is:

(5)

For FFT-MAG, the network output is , and we have theestimation error as:

(6)

In practice, when speech is present, hence we assume. When (overestimation),

, which is exponential with respect to . In this case,clearly grows much faster than , which is linear withrespect to . For the case when ,and . To see which error is greater,we plot in Fig. 5(a) the contour of the ratio betweenand by varying and . We can see that isconsistently smaller than except only when both andare very small (the region where the ratio is less than 1). This

can also be seen in Fig. 5(b), where we plot normalized errorcurves ( and ) when .The analysis implies that, when using supervised techniques

for estimating the clean magnitude, which is the underlyinggoal, one is likely more accurate by predicting the maskingfunction FFT-MASK rather than the log magnitude. Roughlyspeaking, this is because the errors are magnified by the ex-ponential expansion when converting to the magnitude domainbefore resynthesis. This analysis does not depend on the fre-quency scale, hence it also applies to the Gammatone frequency

scale, partly explaining why predicting GF-POW is worse thanpredicting the IRM. In addition, we can show that similar anal-ysis applies to other types of nonlinear compression, such as thecubic root compression , and even to masks involving non-linear compression, such as the LOGMASK mentioned above.The above analysis assumes that the network learns equallywell for both FFT-MAG and FFT-MASK, i.e. assuming thesame . Actually, training on FFT-MAG seems harder than onFFT-MASK, as indicated by the consistently larger valuesacross frequency, shown in Fig. 4. This is perhaps due to thespeaker-independence setting, where the patterns in FFT-MAGare more sensitive to speaker variations.The log or root compression is often used for reducing the dy-

namic range. If one wants to predict the clean magnitude, sucha compression is also needed in neural networks to avoid nu-merical issues so that gradients can flow well. Nevertheless, weargue that regardless of the frequency scale, such a compressionis better achieved via masking, which can be thought as a formof normalization, instead of a nonlinear compression.

C. Comparison with Other Systems

We now compare with a very recent supervised NMF system[31], i.e. ASNA-NMF, and a recent speech enhancement system[13], which we call SPEH. The results of these two systemsare shown in the bottom of Tables II, III, and IV. ASNA-NMFis also a data driven method, trained on the same data as usedby the DNNs. Its performance, however, is significantly worsethan supervised speech separation. The performance differencein terms of STOI is particularly large compared to our DNNthat predicts the IRM, e.g. 10% worse for the dB enginenoise. In challenging cases, e.g. dB babble, ASNA-NMFimproves upon unprocessed mixtures by only 2 percentagepoints in STOI. ASNA-NMF on average achieves significantlyworse STOI but better PESQ results compared to binarymasking (IBM and TBM). However, its PESQ results areconsistently worse than ratio masking (IRM and FFT-MASK).Informal listening indicates that the output from ASNA-NMFhas noticeable speech distortions and residual noise even at5 dB. Speech enhancement, which does not rely on training,seems to have difficulty in challenging test conditions. SPEHfails to improve STOI for 4 out of 5 noises at dB, and 3out of 5 noises at 0 dB. Even at 5 dB where spectral patterns of

Page 9: IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND ...web.cse.ohio-state.edu/~wang.77/papers/WNW.taslp14.pdf1850 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22,

WANG et al.: ON TRAINING TARGETS FOR SUPERVISED SPEECH SEPARATION 1857

TABLE VGENERALIZATION PERFORMANCE ON TWO UNSEEN NOISES AT dB

TABLE VIGENERALIZATION PERFORMANCE ON TWO UNSEEN NOISES AT 0 DB

TABLE VIIGENERALIZATION PERFORMANCE ON TWO UNSEEN NOISES AT 5 DB

speech are prominent, SPEH is outperformed by both DNN andASNA-NMF. This is to be expected as the latter techniques aredata-driven. With the same ground truth signal, FFT-MASKsignificantly outperforms ASNA-NMF and SPEH in termsof SNR. For example, at dB, the average output SNR ofFFT-MASK is 2.13 dB and 3.41 dB better than ASNA-NMFand SPEH, respectively.In practice, supervised speech separation is often trained on

multiple noises (i.e., multi-condition training) for good gener-alization (e.g. [36]). We train such a system on all 5 noises topredict the IRM. This system is called MC-IRM and its perfor-mance is also shown in Table II to IV. We can see that on av-erage the performance does not degrade thanks to the represen-tational power of DNNs. In fact, the performance of multi-con-dition training even improves over individually trained modelssometimes (e.g., for dB factory1), and the performance ad-vantage over ASNA-NMF and SPEH remains.We further compare the generalization performance on

two unseen noises–a different factory noise (factory2) anda tank noise, both from NOISEX. We compare MC-IRMwith ASNA-NMF and SPEH. The two data driven systemsMC-IRM and ASNA-NMF are both trained on the previouslyused 5 noises at -5 and 0 dB. Note that the main purpose ofthis set of experiments is to compare relative performance. Aswe are training on only 5 noises, the presented results by nomeans represent the best obtainable ones for either MC-IRMor ASNA-NMF. It is expected that the performance will besignificantly improved using more noises for training, at leastfor MC-IRM as indicated by the results in [36], [39]. Thegeneralization results at , 0, and 5 dB are shown in Tables V,Tables VI, and Tables VII, respectively. Note that at 5 dB,the SNR, speakers and noises are all new. We can see thatMC-IRM again outperforms the other two systems on bothnoises across all SNR conditions, especially in terms of STOI.

The STOI and PESQ improvements of MC-IRM over unpro-cessed mixtures are large, while the STOI improvements of theother two systems are limited or marginal. We should also pointout that with more training data, separation in the test phasetakes significantly more time for ASNA-NMF. In contrast, thisdoes not affect supervised speech separation systems, wherethe additional computational burden occurs only in the trainingphase.

VI. CONCLUDING REMARKS

Choosing a suitable training target is critical for supervisedlearning, as it is directly related to the underlying computationalgoal. In the context of speech separation, the speech resynthe-sized (either directly or indirectly) using any ideal target restoresintelligibility and quality. In practice, however, since the targetshave to be estimated, the choice should be made by consideringhow well it can be estimated and how the errors in estimationaffect performance.Traditionally, the IBM is used as the training target for su-

pervised speech separation. Despite the simplicity of the IBMand the recent success of classification based speech separa-tion it has inspired, it is unclear whether the IBM is the besttarget in terms of estimation. In this study, we have system-atically investigated the relative performance between varioustraining targets, some new and others not, using both objectiveintelligibility and quality metrics. The compared targets can becategorized into binary masking based (IBM and TBM), ratiomasking based (IRM and FFT-MASK), and spectral envelopebased (FFT-MAG and GF-POW) targets. In general, we havefound that binary masking produces worse objective qualityresults compared to ratio masking. We also found that binarymasking leads to slightly worse objective intelligibility resultsthan ratio masking. This is likely because predicting ratio tar-gets is less sensitive to estimation errors than predicting bi-nary targets. An unexpected finding of this study is that the di-rect prediction of spectral envelopes produces the worst results,as best illustrated by the substantial performance gap betweenFFT-MAG and FFT-MASK, where the two targets are essen-tially two alternative views of the same underlying goal, theclean speech magnitude. Aside from the analysis presented inSectionV-B, which points to the issue of nonlinear compression,we believe that masking has several advantages over spectralenvelope estimation. Perhaps most importantly, masks make di-rect contact with the observed mixtures in the sense that they areused to modulate the mixtures in the time-frequency domain. Incontrast, direct estimation of speech magnitude ignores the mix-ture which contains the true underlying signal. This can be prob-lematic when there is a significant amount of erroneous estima-tion. Furthermore, ideal masks are inherently normalized andusually bounded, potentially making training easier and predic-tion more accurate compared to unbounded spectral envelopebased targets. Finally, ideal masks are likely easier to learn thanspectral envelopes, as their spectrotemporal patterns are morestable with respect to speaker variations.We have also compared a supervised IRM estimation algo-

rithm with recent algorithms in supervised NMF and statisticalmodel-based speech enhancement. The comparisons across var-ious test conditions clearly indicate significant performance ad-vantage of our system.

Page 10: IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND ...web.cse.ohio-state.edu/~wang.77/papers/WNW.taslp14.pdf1850 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22,

1858 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 22, NO. 12, DECEMBER 2014

To conclude, we hope efforts will be devoted to the designof new training targets in the future, which has the potential tofurther improve performance without adding significant com-putational burden. For example, predicting intermediate targetsthat encode more structure [37] or are easier to learn [9] hasbeen shown to be useful.

REFERENCES[1] M. Anzalone, L. Calandruccio, K. Doherty, and L. Carney, “Determi-

nation of the potential benefit of time-frequency gain manipulation,”Ear Hear., vol. 27, no. 5, pp. 480–492, 2006.

[2] D. Brungart, P. Chang, B. Simpson, and D. Wang, “Isolating the en-ergetic component of speech-on-speech masking with ideal time-fre-quency segregation,” J. Acoust. Soc. Amer., vol. 120, pp. 4007–4018,2006.

[3] C. Chen and J. Bilmes, “MVA processing of speech features,” IEEETrans. Audio, Speech, Lang. Process., vol. 15, no. 1, pp. 257–270, Jan.2007.

[4] J. Chen, Y. Wang, and D. Wang, “A feature study for classification-based speech separation at very low signal-to-noise ratio,” in Proc.ICASSP, 2014, pp. 7059–7063.

[5] J. Duchi, E. Hazan, and Y. Singer, “Adaptive subgradient methods foronline learning and stochastic optimization,” J. Mach. Learn. Res., pp.2121–2159, 2011.

[6] J. Erkelens, R. Hendriks, R. Heusdens, and J. Jensen, “Minimummean-square error estimation of discrete fourier coefficients with generalizedgamma priors,” IEEE Trans. Audio, Speech, Lang. Process., vol. 15,no. 6, pp. 1741–1752, Aug. 2007.

[7] J. Garofolo, DARPA TIMIT acoustic-phonetic continuous speechcorpus. Gaithersburg, MD, USA: Nat. Inst. of Standards Technol.,1993.

[8] X. Glorot, A. Bordes, and Y. Bengio, “Deep sparse rectifier networks,”in Proc. 14th Int. Conf. Artif. Intell. Statist. JMLR W&CP Volume,2011, vol. 15, pp. 315–323.

[9] C. Gulcehre and Y. Bengio, “Knowledge matters: Importance of priorinformation for optimization,” in Proc. Int. Conf. Learn. Representat.(ICLR), 2013.

[10] K. Han and D. Wang, “A classification based approach to speech seg-regation,” J. Acoust. Soc. Amer., vol. 132, pp. 3475–3483, 2012.

[11] K. Han, Y.Wang, and D.Wang, “Learning spectral mapping for speechdereverberation,” in Proc. ICASSP, 2014, pp. 4648–4652.

[12] E. Healy, S. Yoho, Y. Wang, and D. Wang, “An algorithm to improvespeech recognition in noise for hearing-impaired listeners,” J. Acous.Soc. Amer., pp. 3029–3038, 2013.

[13] R. Hendriks, R. Heusdens, and J. Jensen, “MMSE based noise PSDtracking with low complexity,” in Proc. ICASSP, 2010, pp. 4266–4269.

[14] J. R. Hershey, S. J. Rennie, P. A. Olsen, and T. T. Kristjansson,“Super-human multi-talker speech recognition: A graphical modelingapproach,” Comput. Speech Lang., pp. 45–66, 2010.

[15] G. E. Hinton, N. Srivastava, A. Krizhevsky, I. Sutskever, and R. R.Salakhutdinov, “Improving neural networks by preventing co-adapta-tion of feature detectors,” arXiv preprint arXiv:1207.0580, 2012.

[16] Z. Jin and D. Wang, “A supervised learning approach to monaural seg-regation of reverberant speech,” IEEE Trans. Audio, Speech, Lang.Process., vol. 17, no. 4, pp. 625–638, May 2009.

[17] G. Kim, Y. Lu, Y. Hu, and P. Loizou, “An algorithm that improvesspeech intelligibility in noise for normal-hearing listeners,” J. Acoust.Soc. Amer., pp. 1486–1494, 2009.

[18] U. Kjems, J. Boldt, M. Pedersen, T. Lunner, and D. Wang, “Role ofmask pattern in intelligibility of ideal binary-masked noisy speech,” J.Acoust. Soc. Amer., vol. 126, pp. 1415–1426, 2009.

[19] N. Li and P. Loizou, “Factors influencing intelligibility of ideal bi-nary-masked speech: Implications for noise reduction,” J. Acoust. Soc.Amer., vol. 123, no. 3, pp. 1673–1682, 2008.

[20] Y. Li and D. Wang, “On the optimality of ideal binary time–frequencymasks,” Speech Commun., pp. 230–239, 2009.

[21] P. C. Loizou, Speech Enhancement: Theory and Practice. BocaRaton, FL, USA: CRC, 2007.

[22] N. Mohammadiha, P. Smaragdis, and A. Leijon, “Supervised and un-supervised speech enhancement approaches using nonnegative matrixfactorization,” IEEE Trans. Audio, Speech, Lang. Process., vol. 21, no.10, pp. 2140–2151, Oct. 2013.

[23] A. Narayanan and D. Wang, “Ideal ratio mask estimation using deepneural networks for robust speech recognition,” inProc. ICASSP, 2013,pp. 7092–7096.

[24] A. Narayanan and D. Wang, “The role of binary mask patterns in auto-matic speech recognition in background noise,” J. Acoust. Soc. Amer.,pp. 3083–3093, 2013.

[25] R. Plomp, The Intelligent Ear: On the Nature of Sound Perception.Mahwah, NJ, USA: Lawrence Erlbaum Associates, 2002.

[26] A. M. Reddy and B. Raj, “Soft mask methods for single-channelspeaker separation,” IEEE Trans. Audio, Speech, Lang. Process., vol.15, no. 6, pp. 1766–1776, Aug. 2007.

[27] A. Rix, J. Beerends, M. Hollier, and A. Hekstra, “Perceptual evalua-tion of speech quality (PESQ)-a new method for speech quality assess-ment of telephone networks and codecs,” in Proc. ICASSP, 2001, pp.749–752.

[28] S. Srinivasan, N. Roman, and D. Wang, “Binary and ratio time-fre-quency masks for robust speech recognition,” Speech Commun., vol.48, no. 11, pp. 1486–1501, 2006.

[29] C. Taal, R. Hendriks, R. Heusdens, and J. Jensen, “An algorithm forintelligibility prediction of time-frequency weighted noisy speech,”IEEE Trans. Audio, Speech, Lang. Process., vol. 19, no. 7, pp.2125–2136, Sep. 2011.

[30] A. Varga and H. Steeneken, “Assessment for automatic speech recogni-tion: II. NOISEX-92: A database and an experiment to study the effectof additive noise on speech recognition systems,” Speech Commun.,vol. 12, pp. 247–251, 1993.

[31] T. Virtanen, J. Gemmeke, and B. Raj, “Active-set Newton algorithmfor overcomplete non-negative representations of audio,” IEEE Trans.Audio, Speech, Lang. Process., vol. 21, no. 11, pp. 2277–2289, Nov.2013.

[32] D. Wang, “On ideal binary mask as the computational goal of auditoryscene analysis,” in Speech Separation by Humans and Machines, P.Divenyi, Ed. Norwell, MA, USA: Kluwer, 2005, pp. 181–197.

[33] D. Wang, U. Kjems, M. Pedersen, J. Boldt, and T. Lunner, “Speechintelligibility in background noise with ideal binary time-frequencymasking,” J. Acoust. Soc. Amer., vol. 125, pp. 2336–2347, 2009.

[34] Y. Wang, K. Han, and D. Wang, “Exploring monaural features forclassification-based speech segregation,” IEEE Trans. Audio, Speech,Lang. Process., vol. 21, no. 2, pp. 270–279, Feb. 2013.

[35] Y. Wang and D. Wang, “Cocktail party processing via structured pre-diction,” in Proc. NIPS, 2012, pp. 224–232.

[36] Y. Wang and D. Wang, “Towards scaling up classification-basedspeech separation,” IEEE Trans. Audio, Speech, Lang. Process., vol.21, no. 7, pp. 1381–1390, Jul. 2013.

[37] Y. Wang and D. Wang, “A structure-preserving training target for su-pervised speech separation,” in Proc. ICASSP, 2014, pp. 6127–6131.

[38] J. Woodruff, “Integrating monaural and binaural cues for sound local-ization and segregation in reverberant environments,” Ph.D. disserta-tion, The Ohio State Univ., Columbus, OH, USA, 2012.

[39] Y. Xu, J. Du, L. Dai, and C. Lee, “An experimental study on speechenhancement based on deep neural networks,” IEEE Signal ProcessingLett., vol. 21, no. 1, pp. 66–68, Jan. 2014.

Yuxuan Wang received his B.E. degree in networkengineering from Nanjing University of Posts andTelecommunications, Nanjing, China, in 2009.He is currently pursuing his Ph.D. degree at TheOhio State University. He is interested in speechseparation, robust automatic speech recognition, andmachine learning.

Arun Narayanan, photograph and biography not provide at the time ofpublication.

DeLiang Wang, photograph and biography not provide at the time ofpublication.