Effectiveness of dereverberation, feature transformation, discriminative training methods, and system combination approach for various reverberant environments

The recently released REverberant Voice Enhancement and Recognition Benchmark (REVERB) challenge includes a reverberant automatic speech recognition (ASR) task. This paper describes our proposed system based on multi-channel speech enhancement preprocessing and state-of-the-art ASR techniques. For preprocessing, we propose a single-channel dereverberation method with reverberation time estimation, which is combined with multichannel beamforming that enhances direct sound compared with the reflected sound. In addition, this paper also focuses on state-of-the-art ASR techniques such as discriminative training of acoustic models including the Gaussian mixture model, subspace Gaussian mixture model, and deep neural networks, as well as various feature transformation techniques. Although, for the REVERB challenge, it is necessary to handle various acoustic environments, a single ASR system tends to be overly tuned for a specific environment, which degrades the performance in the mismatch environments. To overcome this mismatch problem with a single ASR system, we use a system combination approach using multiple ASR systems with different features and different model types because a combination of various systems that have different error patterns is beneficial. In particular, we use our discriminative training technique for system combination that achieves better generalization by making systems complementary with the modified discriminative criteria. Experiments show the effectiveness of these approaches, reaching 6.76 and 18.60 % word error rates on the REVERB simulated and real test sets. These are 68.8 and 61.5 % relative improvements over the baseline.


Introduction
Automatic speech recognition (ASR) using distant microphones can overcome application restrictions of places and devices and widen the usage of speech interfaces. For example, users can control distant home appliances by voice without touching the devices. However, in such a scenario, it is necessary to address reverberation, which is composed of reflected sounds from walls, ceilings, or furniture, in addition to the direct sound from a sound source. Reverberation as well as noise degrades the intelligibility of speech for humans, and it also significantly degrades ASR performance.
The REverberant Voice Enhancement and Recognition Benchmark (REVERB) challenge is an Audio and Acoustic Signal Processing (AASP) challenge sponsored by the IEEE Signal Processing Society in 2013, and has recently been released for studying reverberant speech enhancement and recognition techniques [1]. This paper focuses on the speech recognition task, which is a mediumsized vocabulary continuous speech recognition task, in order to evaluate the ASR performance in reverberant environments.
In such a scenario, speech enhancement before ASR is important and impacts ASR performance. We have proposed a single-channel dereverberation method [2]. This method first estimates a reverberation time, which is one of the most important parameters for characterizing the extent of reverberation, and attempts to eliminate the reverberant components based on the estimated reverberation time. In addition, in order to exploit the eightchannel data provided by the REVERB challenge, we use a beamforming (BF) approach [3] with a direction-of-arrival estimation [4,5].
In addition to the speech enhancement process, we focus on the state-of-the-art ASR techniques. Recently, ASR performance has been significantly improved owing to various types of discriminative training [6,7] and feature transformations [8][9][10][11][12][13]. In the previous Computational Hearing in Multisource Environments (CHiME) challenge [14], we showed the effectiveness of discriminative training and feature transformations in noisy environments [15,16], and this time, also our proposed system employs these techniques. However, the CHiME challenge and other existing evaluation campaigns for noise-robust ASR [14,17] mainly focus on the variety of non-stationary additive noises, and the variety of room shapes or room types in these campaigns is very limited. On the other hand, the REVERB challenge [1] includes eight different reverberant environments: four rooms, which are composed of three simulated rooms and one real recorded room, multiplied by two types of source-to-microphone distances. In this scenario, due to the variety in the evaluation environments and the mismatch between simulated training data and real test data, discriminative training would cause over-training problems, although discriminative training is very powerful for matched conditions where training and evaluation conditions are close, in general. Therefore, it is important to confirm that speech recognition systems with discriminative training and feature transformations perform robustly in various reverberant environments.
This paper deals with two feature transformation approaches: linear transformation and non-linear discriminative feature transformation. The former approach converts original feature vectors to new feature vectors based on linear transformation matrices. This paper deals with linear discriminant analysis (LDA) [8] and maximum likelihood linear transformation (MLLT) [9,10] to estimate the transformation matrices. LDA uses long context input features, which are obtained by concatenating multiple features in contiguous frames, as original feature vectors to exploit feature dynamics. Therefore, LDA can reduce the influence of reverberation because the long context input features can handle the distorted speech features across several frames due to the influence of longer reverberation than the window size of the shorttime Fourier transform (STFT) [18,19]. This property is particularly effective for reverberant speech recognition, and this paper investigates the effectiveness of LDA on ASR performance in detail with varying context sizes. In addition, MLLT finds a linear transformation of features to reduce state-conditional feature correlations. For the latter approach, we use non-linear discriminative feature transformation [12], which directly reduces ASR errors by estimating non-linear feature transformation matrix with discriminative criteria.
The above feature transformation techniques estimate transformation matrices in the training stage. However, to improve recognition accuracy for unknown conditions in the evaluation stage, the adaptation strategy of estimating feature transformation matrices for evaluation data is also effective. This paper deals with basis feature-space maximum likelihood linear regression (basis fMLLR) [20], which can estimate transformation matrices robustly even in the cases of short utterances. In addition, in the training stage, speaker adaptive training (SAT) [11] is also used. It trains acoustic models in a canonical speaker space based on the MLLR framework in order to obtain better feature transformation in the adaptation stage.
After the feature transformations, Gaussian mixture model (GMM)-based acoustic models are obtained by using discriminative training techniques [6,7] and also this paper deals with deep neural networks (DNN) [13] that have recently attracted great attention, and we have shown promising results in noisy environments [16]. Note that the lower layers of a DNN play the role of discriminative feature transformation [21], and our DNN system skips discriminative feature transformation, which is already included in a DNN.
The studies above mainly focus on a single ASR system. On the other hand, the use of multiple systems is another solution to improve the robustness of ASR performance [22][23][24]. For our proposed method, which exploits discriminative training methods, the best performing system is different from environment to environment due to the variety of evaluation data or mismatch between training and evaluation data. The system combination methods relax the degradation of speech recognition performance coming from these varieties or mismatches, e.g., [25,26] proposed to use a complementary system for system combination. This paper constructs various systems that have different properties, and in particular, our proposed discriminative training method introduces complementary systems intentionally within a lattice-based discriminative training framework [27,28]. The results from various recognizers will be combined using recognizer output voting error reduction (ROVER) [22].
In summary, there are three objectives in this paper: First, the effectiveness of dereverberation and microphone-array speech enhancement techniques is validated. Second, the effectiveness of feature transformation and discriminative training for reverberant environments is validated. The objectives here are various types of acoustic modeling such as the GMM, subspace Gaussian mixture model (SGMM) [29], and DNN and their discriminative training. Third, to address the variety of reverberant environments, a system combination approach is introduced and its effectiveness is validated.
There are two main differences between this paper and the REVERB challenge workshop paper [30]: First, we add detailed descriptions about validated techniques and the experimental setup. For example, we detail the speech enhancement, feature transformation, and speaker adaptation parts. Second, we compare our proposed method with other participants' systems that were submitted to the workshop, which clarifies the effectiveness of our proposed method. Figure 1 shows a schematic diagram of the proposed system, which consists of three components. The first component is based on a speech enhancement step, which is described in Section 3. This paper focuses on singleand eight-channel data. The speech enhancement part consists of (1) a multichannel delay-and-sum BF with direction-of-arrival estimation that enhances the direct sound compared with the reflected sound, (2) a singlechannel dereverberation technique with reverberation time estimation that attempts to eliminate late reverberation, and (3) a normalized least-mean-squares (NLMS) adaptive filter algorithm that attempts to eliminate shortterm distortions such as microphone difference or speech distortions caused by speech enhancement methods.

System overview
The second component is based on a feature transformation step, including several feature-level transformations (LDA, MLLT, and basis fMLLR) and discriminative feature transformation (Section 4.1). This step uses two types of features [Mel-frequency cepstral coefficients (MFCC) and perceptual linear prediction (PLP)]. By using two different types of features, it is believed that complementary hypotheses can be obtained for system combination.
The third component is based on the ASR decoding step that uses a discriminatively trained acoustic model with margin control. Three types of systems (GMM, SGMM, and DNN) are constructed. Boosted maximum mutual information (bMMI) is used for GMM and SGMM in Sections 4.2 and for DNN in Section 4.4.
In addition, Section 4.5 describes our proposed system combination approach that combines discriminatively trained complementary systems. In addition to the three types of SAT model, a GMM model without SAT is also constructed; our proposed method constructed

Speech enhancement
This section deals with speech enhancement methods: delay-and-sum BF with cross-spectrum phase (CSP) analysis in Section 3.1, a proposed dereverberation method in Section 3.2, and an NLMS algorithm that attempts to eliminate short-term distortion in Section 3.3. We describe them step by step. The delay-and-sum BF using the CSP method and NLMS adaptive filter algorithm is used for an 8-channel (ch) system; the dereverberation method is used for both the 1-ch and 8-ch systems.

Delay-and-sum BF after direction-of-arrival estimation using CSP method
To enhance the direct sound from the source, a frequencydomain delay-and-sum BF is applied [3]. The timedomain sth sample z m (s) observed by the mth microphone is transformed into the STFT spectrum. The spectrum x t,m (n) at the tth frame and nth frequency bins obtained as where ϕ is a frame shift, and φ is a window function with the window length N F . A vector form of the spectrum x t,m denotes [ x t,m (0), . . . , x t,m (N F − 1)] ∈ C N F , where denotes a transpose of vectors or matrices. The enhanced spectrumx t (n) is obtained by summing the spectrum x t,m (n) with a compensation of a time delay as The arrival time delay τ t,m of the mth microphone from the first microphone is related to the direction of arrival at the tth frame (here τ t,1 = 0). This time delay is estimated by CSP analysis [4]. First, an inverse STFT transform a cross-power spectrum between first and mth microphones into the time domain as where "*" denotes a complex conjugate. The highest correlated point is the maximum point of elements among {ψ t,m (0), . . . , ψ t,m (N F − 1)}. Thus, the time delay τ t,m is calculated as where f samp is a sampling frequency. To improve the performance of the original CSP method, we used a peakhold process [31] and noise component suppression, which sets the cross-power spectrum to zero when the estimated signal-to-noise ratio (SNR) is below 0 dB [5]. Synchronous addition of multiple microphone pair-wise CSP coefficients reduces the noise influence [32].

Single-channel dereverberation with estimation of reverberation time
For a single-channel dereverberation method, we employ an algorithm proposed in [2]. The proposed algorithm is briefly described below, and detailed discussions are found in [2]. Since the proposed method is independently processed across microphones, we omit the microphone index m. When reverberation time T r is much longer than the frame size, an observed power spectrum is modeled as a weighted sum of the source's power spectrumX t ∈ R N F . The source's power spectrum is estimated as follows in the existence of stationary noise N ∈ R N F when the spectrum between frequency bins is independent: where μ and w are the delay frame and the weight coefficient, respectively. The source's power spectrumX t is related to X t aŝ where η is the ratio of a direct sound component to the sum of the direct and reflected sound components, which is a decreasing function of T r because longer T r increases the energy of the reflected sound components.
Here, we assume that the reverberation time T r and η are independent of frequency bins, for simplicity. Assuming that w 0 is unity to normalize reverberation decay for the direct sound, Eq. (7) can be derived from the above relations: Reverberation is divided into two stages: early reverberation and late reverberation. The threshold between them is denoted by D (frames) after the arrival of a direct sound.
Generally, late reverberation mainly degrades speech recognition performance and early reverberation can be ignored. Therefore, the proposed method only focuses on late reverberation. Early reverberation is complex because it greatly depends on room shapes and distributions of room materials, whereas late reverberation is statistical and the sound-energy density decays exponentially with time under the assumption of a diffuse sound field. These are modeled according to Polack's statistical model [33], and w μ is determined as which corresponds to a reverberation decay in Fig. 2.
Here, α s is a subtraction parameter to be set. The upper condition and lower condition correspond of Eq. (8) to the early and late reverberations, respectively. Assuming η is constant, Eq. (7) is a process similar to spectral subtraction [34]. If the subtracted power spectrumX t is less than βX t , it is substituted with βX t . This process is called a flooring, and β is a flooring parameter. We define the floored ratio ρ as a ratio of the number of floored time-frequency bins to the total number of bins. The proposed method estimates a reverberation time T r from a flooring ration ρ. Two observations are exploited for this estimation. First, when some arbitrary reverberation times (T a ) are assumed, ρ increases monotonically with T a because a longer T a increases the extent of subtraction. This is modeled as a linear relation with the inclination ρ . Second, ρ increases with T r at the same T a . Since actual η(T r ) decreases with T r , the power spectrum after dereverberation assuming a constant η is more likely to be floored for a longer T r because the second term of Eq. (7) is larger than that of the actual one in the condition with a longer T r . Therefore, T r has a positive correlation with ρ . This can be modeled as with two predetermined constants a and b.
The estimation process of T r is summarized as follows: Calculate ρ and the inclination ρ by a least-squares regression for some values of arbitrary assumed reverberation times T a , and estimate an actual reverberation time T r by Eq. (9).

NLMS adaptive filter algorithm
The goal of the NLMS adaptive filter algorithm is to eliminate short-term distortions from an observed distorted signal sequence z s =[ z(s−N L +1), . . . , z(s)] ∈ R N L based on a desired signal d s [35] by using a linear filter with the tap length N L . Filters w s ∈ R N L that realize these requirements are recursively trained in a manner where errors between filtered signals and desired signals are minimized as An LMS algorithm uses instantaneous values for the estimation of a gradient, and an NLMS algorithm normalizes the step size parameter by the signal power. Thus, the update formula of an NLMS algorithm is obtained as where is a step size, and is a very small constant that avoids the instability of the update formula. The initial value of filter w 0 is 0. In this case, z s is a reverberant speech, and d s is a clean speech without reverberation. A filter w is obtained from the entire training data set. For evaluation, desired signals d s cannot be obtained; thus, the filter cannot be changed. The tap length of NLMS is short because the goal of this filter is to eliminate a shortterm distortion, whereas the proposed dereverberation algorithm (3.2) attempts to eliminate late reverberation.

Feature transformation and speaker adaptation
Static features concatenated during the left L frames, current frame, and the right R frames are compressed into low-dimensional (I -dimensional) features by using LDA. The class of LDA is the state of the triphone HMM. In addition to this, to reduce the correlation between feature dimensions, MLLT is used. Combined feature transformation is realized as where y t is an original I-dimensional feature at the tth frame, and y t is an I -dimensional transformed feature; For adaptation, instead of a normal fMLLR transformation, the basis fMLLR [20] is used. It can robustly estimate transform matrices and bias terms even for short utterances. This method realizes the transformation of original features y t into adapted features y t by using pre-trained bases of transform matrices and bias terms and estimating their weights as where A f ν ∈ R I ×I and b f ν ∈ R I are the νth pre-trained basis of an fMLLR transform matrix and bias term, respectively, which are estimated from entire training data. For evaluation, only their weights π ν are estimated.
Moreover, to address the wide variety between speakers, SAT as an acoustic model adaptation [11] is frequently used. In SAT training, acoustic models are trained on speaker-adapted training data, which are transformed into canonical speaker space by using speaker adaptation techniques, in this case, fMLLR. This can reduce the influence of a speaker variation. This paper validates the effectiveness of feature transformations (LDA and MLLT) and adaptation techniques (basis fMLLR and SAT).

MMI discriminative training of acoustic model
MMI discriminative training is a supervised training algorithm that maximizes the mutual information between correct labels and recognition hypotheses. This paper focuses on bMMI [36], where a boosting factor b ≥ 0 is used to introduce a weight depending on phoneme accuracies. The objective function is given as where y r = y 0 , . . . , y T(r)−1 is the rth utterance's feature sequence and T(r) is the total frame number of the rth utterance. The acoustic model parameters λ are optimized by the extended Baum-Welch algorithm. λ is a mean, variance, and mixture weight of GMM. H s r and H s are the HMM sequences of the correct label s r and a hypothesis s, respectively; p λ is the acoustic model likelihood; κ is the acoustic scale; p L is the language model likelihood; and A(s, s r ) is the phoneme accuracy of s for s r . This paper compares the performances of bMMI training of GMM and SGMM to those of maximum likelihood (ML) training.

Discriminative feature transforms
The extension of a discriminative training to a feature transformation is referred to as a feature-space discriminative training [12]. It estimates a matrix M ∈ R I ×J that projects rich, high-dimensional features h t ∈ R J (J I ) down to low-dimensional transformed features, as follows: Usually, Gaussian posteriors of an N g -mix universal background model (UBM) are used for h t [37]. The objective function can be obtained simply by replacing y r with the rth utterance's transformed feature sequence The matrices M are optimized by maximizing the objective function F b (M). In this study, we validate the effectiveness of a feature-space bMMI (f-bMMI).

Discriminative training of DNN
In a DNN-HMM hybrid system, sequential discriminative training according to the (b)MMI criterion (14) has been proposed [38] in addition to a usual cross-entropy (CE) training. The DNN provides posterior probabilities for the HMM state j. The acoustic likelihood p θ is replaced by a pseudo likelihood as where p 0 j is the prior probability of a state j calculated from a forced alignment of the training data. For each HMM state, the model θ includes a softmax activation function: where a j is the activation of the jth unit in the output layer. θ is a parameter in weight matrices and bias terms of DNN. These activations are trained discriminatively according to the bMMI criterion. The bMMI objective function is the same as Eq. (14), simply by replacing λ with θ: F b (θ).

Constructing complementary system suitable for system combination
We describe a discriminative method that constructs complementary systems for appropriate system combination [27,28]. Complementary systems are constructed by discriminatively training a model, which begins with an initial model. The proposed discriminative training method for complementary systems is extended from a discriminative training principle. Assuming Q base systems have already been constructed and fixed, the discriminative training objective function F c for building a complementary system is where where M is the set of model parameters of a complementary system to be optimized; that is, λ, M, and θ. α c is a scaling factor. The model parameter M is shared among the original F and the Q base models' F to be optimized. This subtracts an objective function related to the one-best hypothesis of the qth base system, s q , from an objective function related to the correct label s r . The discriminative criterion F is selected as bMMI or f-bMMI. If α c equals zero, this objective function matches the original F. The first term in Eq. (19) promotes a good performance according to the discriminative training criterion, whereas the second term makes the target system generate hypotheses that have different tendencies from the original Q base models. This procedure is commonly used to obtain the objective functions of Sections 4.2, 4.3, and 4.4.

REVERB challenge speech recognition task
We validated the effectiveness of our proposed approaches for a reverberated speech recognition task on the REVERB challenge [1] data. The task is a mediumvocabulary ASR in reverberant environments, whose utterances are taken from the Wall Street Journal (WSJ) database (WSJCAMO [39]). This database includes two types of data: SIMDATA created by convolving clean speech with six types of room impulse responses at a distance of 0.5 m (near) or 2 m (far) from the microphones in three offices (Rooms 1, 2, and 3) whose reverberation times are 0.25, 0.5, and 0.75 s, respectively, with relatively stationary noise at 20 dB SNR; and REALDATA created by recording real-world speech at a distance of 1 m or less (near) or 2.5 m or less (far) from the microphones in one room (Room 1) with stationary noise such as air conditioner noise. Eight microphones were arranged on the circle with a radius of 0.1 m. The number of speakers and utterances of the training set (tr), evaluation set (eva), and development set (dev) is shown in Table 1. Acoustic models were trained using tr. Some of the parameters, e.g., language model weights, were tuned based on the WERs of dev. The vocabulary size is 5 k, and a trigram language model is used. The REVERB challenge speech recognition task is categorized in terms of processing techniques, training data of the acoustic model, recognizer type, and number of channels used, as shown in Table 2. All experiments in this paper were "utterancebased batch processing, " 1 "acoustic model trained on the challenge provided multicondition (MC) training data, " "own recognizer, " and "single-or eight-channel data". These systems were constructed by using the Kaldi toolkit [40].

Speech enhancement
The REVERB challenge provides single-, two-, and eightchannel data. We used single-and eight-channel data. For single-and eight-channel data, the proposed dereverberation technique was used with parameters: D = 9, α = 5, β = 0.05, a = 0.005, and b = 0.6. For eight-channel data, before dereverberation, delay-and-sum BF with a direction of arrival estimation by CSP analysis was performed, which used a total of 8 C 2 (= 28) pairs of microphones. After dereverberation, NLMS adaptive filters with N L = 200 taps were applied.  Italicized data denotes the category to which this paper belongs

Feature extraction and transformation and acoustic model adaptation
We describe the settings of acoustic features and feature transformations, which are detailed in [15,16]. The baseline acoustic features were 0-12 order MFCCs and PLPs with first and second dynamic features. After concatenating static MFCCs/PLPs during L + R + 1 frames without using delta feature, a total of (13×(L+R+1))-dimensional features were compressed into 40 dimensions by the LDA. For adaptation, when speaker IDs were known for the training set, bases A f ν and b f ν were estimated. For the development and evaluation set, speaker IDs are assumed to be unknown, and weight vector π ν was estimated.

Discriminative methods
In discriminative feature transformation (Section 4.3), a UBM with N g = 400-mix Gaussians was used.

Building acoustic models
First, clean acoustic models were trained. The number of monophones was 45, including silence ("sil"). Triphone model has 2500 states and 15,000 Gaussian distributions. Second, using the alignments and triphone tree structures of the clean model, reverberated acoustic models were trained on the MC dataset according to the ML The proposed dereverberation method was used. Three types of acoustic models (GMM, SGMM, and DNN) were constructed with feature transformation (LDA + MLLT), adaptation (basis fMLLR and SAT), and discriminative training (bMMI and f-bMMI). Subscript letter "c" represents the proposed "complementary" system. Italicized data were the best systems in each condition criterion. Finally, from this ML model, we performed the discriminative training and feature transformations. For DNNs, we used Povey's implementation of neural network training in Kaldi [40]. DNN has two hidden layers was two and each hidden layer has 642 nodes. The total number of parameters was 2 M. The initial learning rate of CE training was 0.02, and this decreased to 0.004 at the end of training. The training targets for the DNN were determined by the forced alignments on reverberant speech using a GMM model with SAT. The parameters used in our experiments were set as those in the WSJ tutorial (s6) attached to the Kaldi toolkit, although some settings such as the number of model parameters or some minor parameters were modified.

System combination
We prepared three types of ASR acoustic model systems for the challenge: GMM, SGMM, and DNN. To improve the performance of the respective systems, for GMM, f-bMMI was used; whereas for SGMM and DNN, bMMI was used. On the development set, because output tendencies of GMM with and without SAT model were different, both systems were used for a system combination. For each system, complementary systems were constructed by the proposed method as shown in 4.5. These systems were trained both for MFCC and PLP features; thus, a total of 16 systems were prepared. After decoding for generated lattices, minimum Bayes risk decoding [41], which slightly improved the performance, was commonly used.

Black-box optimization
Bayesian optimization using Gaussian processes [42] was applied to various speech recognition problems including neural network [43] and HMM topology optimization [44]. In this paper, we also applied this technique to the selection of combined systems and the parameter optimization for ROVER. The objective function of the optimization was WER of the development set. Tables 3 and 4 show the WERs of the development set (dev) for three simulated rooms and one real room with two types of source-to-microphone distances (near/far). Table 3 is based on a single-channel one and Table 4 is based on an eight-channel one. The "Kaldi baseline" in Table 3 is an acoustic model trained on the MC data without speech enhancement. "derev." is the proposed dereverberation method with a reverberation time estimation. Although, for some cases in room 1, the reverberation time is fairly short and the proposed method degraded In addition to the proposed dereverberation method, BF with direction of arrival estimation by CSP analysis and NLMS adaptive filters were used. Subscript letter "c" represents the proposed "complementary" system. Italicized data were the best systems in each condition Subscript letter "c" represents the proposed "complementary" system. Italicized data were the best systems in each condition performance, for other cases and on average, performance was improved by approximately 2 %. Weninger et al. [45] showed that our proposed dereverberation technique is effective even with a state-of-the-art denoising auto-encoder. For the eight-channel data shown in Table 4, BF with "derev." significantly improved performance by approximately 6.3-8.3 % on average, because the direction of arrival estimation was stable and reliable. "NLMS" improved the WER by 2.0 % for the REAL-DATA, but degraded the WER by 0.6 % for the SIM-DATA. However, because these decreases in performance have less impact than the improvements, we used NLMS below. These results above used MFCC features. Experimental results using PLP features are shown in Table 5. On average, the ASR performances using PLP features were approximately 0.2-1 % lower than those using MFCC features; however, their error tendencies were fairly different, which was a good property for system combination.

LDA and MLLT feature transformation and adaptation
LDA and MLLT feature transformations significantly improved performance by approximately 2.6-5.5 %. Table 6 shows the effect of an LDA context size on performance. The performance of the SIMDATA could not be improved by context sizes longer than 4. For the REALDATA, performance could be improved in several cases by adding more right context, but generally not by adding left context. In reverberant environments, because reverberant components of current frames give an influence on the features in the right context, the right context can be useful for improving speech recognition performance. In the end, we kept the context size at the default setting, L = R = 4. Tables 3 and 4 show that the adaptation technique, basis fMLLR, improved performance by approximately 1.3-6.9 %. The effect of SAT is unstable between environments.  Italicized data were the best systems in each condition Tables 3 and 4 show that the discriminative training was effective for reverberant environments. The performances of f-bMMI training were higher than those of bMMI training in all cases by approximately 0.6-1.7 %. The WERs of our complementary systems were only slightly lower (0.2-0.7 %) than those of the base systems, and they have different tendencies from base systems; thus, they appear to be well suited to system combination. Table 7 shows the effect of the iteration numbers of bMMI and f-bMMI on the development set performance. The results show that the best performance was achieved at four iterations. Tables 3 and 4 show the performance of SGMM acoustic models. For the SIMDATA, the performance of SGMMs was higher than that of GMMs. However, for the REAL-DATA, the performance was lower than that of GMMs. Because the REALDATA were noisier than the SIMDATA, the estimation of speaker vector can be unstable.

SGMM and DNN
DNN acoustic models achieved the best performance for the SIMDATA. Although the best system for the REAL-DATA was GMM without SAT, DNN was the second best. On average over the SIMDATA and REALDATA, DNNs achieved the best performance. Although DNN was trained discriminatively even by CE training according to the frame-level discriminative criterion, sequence discriminative training, bMMI, for DNN systems turned out to be as effective as for other systems.

System combination
We tested five types of system combinations, as shown in Table 8. The number 2 stands for one MFCC system and one PLP system. The number 4 stands for two MFCC and two PLP systems composed of a base system and the proposed complementary system. These systems' outputs are combined by using ROVER. The ID 1) system was a combination of SAT-GMMs (f-bMMI) using both MFCC and PLP features. The performance for the REALDATA improved by 1.2-4.2 % over the f-bMMI with a SAT (MFCC) single system. For the GMM system without SAT, using f-bMMI [ID 2)], the WER improved by 0.2-1.5 % for the SIMDATA and 0.6-1.4 % for the REALDATA. Including the complementary systems  , the performance was again improved; this system, which combined 16 systems in total, achieved the best average performance on the development set. For the reference, the results of eight system combination without using our proposed combination are added to the last line of 1 ch case [ID 6)]. The WER on REALDATA was worse than those of the proposed 16 system combination, which shows that the complementary training generalizes the ASR results for unseen data conditions more.
In all cases except for the room 1/far(8-ch) condition, 2 the performances were better than those of the best system. This shows that the system combination approach is effective for the case where reverberant environments are various.

Black-box optimization
For eight-channel data, black-box optimization was performed. Figure 3 shows the average WER in terms of the iteration number. WER almost decreased monotonically and, after 100 iterations, it converged. Among these iterations, the results that achieved the best WER on average, are shown in the last column of Table 8. The performance improved mainly for the REALDATA. Table 9 shows the results for the evaluation set (eva). Legend of the table is the same to the development set. The optimal system combination is determined based on the WER on the development set. The discriminative training of acoustic model (bMMI) and feature-space discriminative training (f-bMMI) significantly improved the performance. SGMM was better than GMM because model adaptation was well performed. DNN outperformed GMM and SGMM. The DNN with discriminative training achieved the best performance for the SIMDATA and REALDATA among single systems. This shows the robustness of DNN in unseen conditions. Moreover, system combination [ROVER 5)] improved the WER by 1.0-1.3 % for the SIMDATA and 2.1-2.2 % for the REALDATA. Among system combination systems, the performance of ROVER 5) was better than that of ROVER 6), which used black-box optimization and could be overly tuned on the development set.

Comparison to other participants' results in the REVERB challenge workshop
The results in the previous section were submitted to the REVERB challenge workshop. Figure 4 shows the WERs for the single-channel data of other participants who belong to the same category, which corresponds to all cases except "own dataset" in the training data of the acoustic models in Table 2. Figure 5 shows those for the eight-channel data. For speech enhancement purposes, a long-short-term memory recurrent neural network (LSTM-RNN) was effective [46] ("TUM2" in the figure). Many participants used DNN-based acoustic modeling (e.g., [47] "Nanyang Tech" in the figure). Speaker adaptation of DNN based on the i-vector technique in addition to robust features, also performed well [48] ("INRS Energie" in the figure). We achieved the best performances in both single-and eight-channel cases.

Conclusions
We evaluated the medium-sized vocabulary continuous speech recognition task of the REVERB challenge in order to validate the effectiveness of single-channel dereverberation and multi-channel beamforming techniques and discriminative training of acoustic model and feature transformation in reverberant environments. For speech enhancement, experiments show the effectiveness of dereverberation of the late reverberation components, and beamforming using multiple microphones that enhances direct sounds compared to the reflected sounds.
For speech recognition, we validated the effectiveness of feature transformations and discriminative training. Experiments show that these techniques are effective across various types of reverberation as well as in noisy environments. To improve robustness in eight types of environments, the system combination approach was used. Systems from 2 to 16 were constructed to address the problem where the best performing system was different from environment to environment. System combination improved performance; in almost all cases, the combined system outperformed the best performing single system. Our proposed method to specifically provide desired complementary systems for system combination further improved performance. The best results were submitted to the REVERB challenge workshop, and our results were the best among the challenge participants in the same category, which clarifies the effectiveness of our proposed approach.

Endnotes
1 This allows for multiple decoding passes per utterance, such as for calculating the fMLLR matrix, but decodes each test utterance separately, without taking into account information from other test utterances, or speaker identities. 2 In this case, GMM(f-bMMI) exhibited the best performance (26.25 % WER). 3 Among all the participants, [49] was the best. This is a state-of-the-art system composed of a liner-prediction based dereverberation technique, DNN based acoustic modeling, and rescoring using RNN language model. The main difference from our system was the use of the "own dataset" that can compensate for the mismatches between training data and evaluation data (especially for the REALDATA) and improve the performance.