# Speaker adaptation in the maximum a posteriori framework based on the probabilistic 2-mode analysis of training models

## Abstract

In this article, we describe a speaker adaptation method based on the probabilistic 2-mode analysis of training models. Probabilistic 2-mode analysis is a probabilistic extension of multilinear analysis. We apply probabilistic 2-mode analysis to speaker adaptation by representing each of the hidden Markov model mean vectors of training speakers as a matrix, and derive the speaker adaptation equation in the maximum a posteriori (MAP) framework. The adaptation equation becomes similar to the speaker adaptation equation using the MAP linear regression adaptation. In the experiments, the adapted models based on probabilistic 2-mode analysis showed performance improvement over the adapted models based on Tucker decomposition, which is a representative multilinear decomposition technique, for small amounts of adaptation data while maintaining good performance for large amounts of adaptation data.

## 1 Introduction

In automatic speech recognition (ASR) systems using hidden Markov models (HMMs), mismatches between the training and testing conditions lead to performance degradation. One of such mismatches results from speaker variation. Thus, speaker adaptation techniques are employed to transform a well-trained canonical model (e.g., speaker-independent (SI) HMM) to the target speaker. Speaker adaptation requires fewer adaptation data than needed to build a speaker-dependent (SD) model. Among speaker adaptation techniques, eigenvoice (EV) expresses the model of a new speaker as a linear combination of basis vectors, which are built from the principal component analysis (PCA) of the HMM mean vectors of training speakers.

In a similar approach, speaker adaptation based on tensor analysis using Tucker decomposition was investigated in, where bases were constructed from the multilinear decomposition of a tensor that consisted of the HMM mean vectors of training speakers. In the approach, all the training models were collectively arranged in a third-order tensor (3-D array):

$ℳ R × D × S$
(1)

where the first, second, and third modes (dimensions) were for the mixture component, dimension of the mean vector, and training speaker. In, Tucker decomposition was used to build bases and in the experiments, speaker adaptation using Tucker decomposition showed better performance than eigenvoice and maximum likelihood linear regression (MLLR) adaptation. The improvement seemed to be attributable to the increased number of adaptation parameters and compact bases. Also noticed in was that the increased number of adaptation parameters did not guarantee a good performance when the amount of adaptation data was small (the determination of the proper number of adaptation parameters for given adaptation data is a model-order selection problem). Extending the tensor-based approach, in, the fourth mode for noise was added (so,$ℳ$ became a 4-D array) so that the training models of various speakers and noise conditions were decomposed.

In this article, we describe a speaker adaptation method using probabilistic 2-mode analysis, which is an application of probabilistic tensor analysis (PTA) to the second-order tensor (i.e., matrix); PTA is an application of probabilistic PCA (PPCA) to tensor objects. Using probabilistic 2-mode analysis, we derive bases from training models in a probabilistic framework, and formulate the speaker adaptation equation in the maximum a posteriori (MAP) framework. The speaker adaptation equation based on the probabilistic approach becomes similar to MAP linear regression (MAPLR) adaptation as shown below. The experiments showed that the proposed method further improved the performance of the speaker adaptation based on Tucker decomposition for small amounts of adaptation data.

The rest of this article is organized as follows. Section 2.1 explains some tensor algebra and tensor decomposition. Section 2.3 explains the probabilistic 2-mode analysis of a set of mean vectors of training HMMs. In Section 2.5, the estimation of the prior distribution of the adaptation parameter is described. Section 2.6 describes the speaker adaptation in the MAP framework using the bases and the prior. Section 2.2 describes the speaker adaptation using Tucker decomposition, which is compared with the probabilistic 2-mode analysis-based method. We explain the experiments in Section 3 and conclude the article in Section 4. Some of the notations used in this article are summarized in Table1.

## 2 Methods

### 2.1 Multilinear decomposition

Following the convention of multilinear algebra, we denote vectors, matrices, and tensors by lowercase boldface letters (e.g., M), uppercase boldface letters (e.g., M), and calligraphic letters (e.g.,$ℳ$), respectively, in this article.

A tensor is a multidimensional array, and an N-dimensional array is called the N th-order tensor (or N-way array). The order of a tensor is the number of indices for addressing the tensor; so the order of$ℳ I 1 × I 2 × ⋯ × I N$ is N. Scalar, vector, and matrix are zeroth-, first-, and second-order tensors, respectively. There are three indices for addressing the array in a third-order tensor as depicted in Figure1.

Tensor algebra is performed in terms of matrix and vector representations of tensors; the mode-n flattening (matricization) of tensor$ℳ$, which is denoted as M(n), is obtained by reordering the elements as follows:

$M ( n ) ∈ ℝ I n × ( I 1 × ⋯ × I ( n − 1 ) × I ( n + 1 ) × ⋯ × I N ) .$
(2)

That is, all the column vectors along the mode n are arranged into a matrix. For example, a third-order tensor$ℳ I × J × K$ can be flattened into an I×(J K), J×(K I), or K×(I J) matrix as depicted in Figure2; for a$ℳ 2 × 2 × 2$ tensor:

$ℳ = M ( : , : , 1 ) M ( : , : , 2 ) M ( : , : , 1 ) = m 111 m 121 m 211 m 221 , M ( : , : , 2 ) = m 112 m 122 m 212 m 222 ,$
(3)

the mode-n flattening is given as: The operation of the mode-n flattening will be denoted as mat n (·), i.e.,$mat n (ℳ)= M ( n )$.

Multiplication of a tensor and a matrix is performed by n-mode product; the n-mode product of a tensor$W$ with a matrix U is denoted as

$ℳ = W × n U$
(5)

and is carried out by matrix multiplication in terms of flattened matrices:

$M ( n ) = U W ( n )$
(6)

or elementwise

$W × n U i 1 … i n − 1 j i n + 1 … i N = ∑ i n = 1 I n w i 1 i 2 … i N u j i n$
(7)

where w and u denote the elements of$W$ and U, respectively. If$W∈ ℝ I 1 × I 2 × ⋯ × I N$ and$U T ∈ ℝ K n × I n$, then the dimension of$W × n U T$ becomes I1×I2××In−1×K n ×In+1××I N .

As an extension of singular value decomposition (SVD) to tensor objects, Tucker decomposition decomposes a tensor as follows:

$ℳ I 1 × I 2 × ⋯ × I N ≃ W K 1 × K 2 × ⋯ × K N ∏ n = 1 N × n U n$
(8)

where$U n ∈ ℝ I n × K n$, K n I n (n=1,…,N). The core tensor$W$ and mode matrices U n ’s correspond to the matrices of singular values and orthonormal basis vectors in matrix SVD, respectively. An example of Tucker decomposition of a third-order tensor is illustrated in Figure3.

The core tensor$W$ and mode matrices U n ’s in Tucker decomposition can be computed such that they minimize

$Error = ∥ ℳ − W ∏ n = 1 N × n U n ∥ 2$
(9)

where the norm of a tensor is defined as$∥X∥= ∑ i 1 = 1 I 1 ∑ i 2 = 1 I 2 … ∑ i N = 1 I N x i 1 i 2 … i N 2$. A representative technique for Tucker decomposition is the alternating least squares (ALS); the basic idea is to compute each mode matrix U n alternatingly with other mode matrices fixed. For more details on Tucker decomposition, refer to. In the following section, we explain probabilistic 2-mode analysis in the context of speaker adaptation.

### 2.2 Speaker adaptation using Tucker decomposition

The probabilistic 2-mode analysis based method is a probabilistic extension of the Tucker decomposition based method. Thus, we compare the probabilistic approach with the Tucker decomposition based method in the experiments. In this section, we explain the speaker adaptation based on the Tucker decomposition of training models in. In this article, speaker adaptation is performed by updating the mean vectors of the output distribution of an HMM. The HMM mean vectors of each training speaker are arranged in an R×D matrix:

$M s = μ s ; 1 … μ s ; r … μ s ; R T , s = 1 , … , S.$
(10)

Here, μs;r denotes the mean vector corresponding to mixture r of the s th training speaker model.

All the centered HMM mean vectors of training speakers,$M s − M ¯ s = 1 S$ where$M ¯ =1/S ∑ s M s$, are collectively expressed as a third-order tensor$ℳ ~$, and we decompose the training tensor by Tucker decomposition as follows:

$ℳ ~ R × D × S ≃ G K R × K D × K S × 1 U mixture × 2 U dim × 3 U speaker = G K R × K D × K S × 3 U speaker × 1 U mixture × 2 U dim .$
(11)

In the above equation,$U mixture ∈ ℝ R × K R$,$U dim ∈ ℝ D × K D$, and$U speaker ∈ ℝ S × K S$ are basis matrices for the mixture component, dimension of the mean vector, and training speaker, respectively (K R R−1, K D D−1, and K S S−1); the core tensor$G$ is common across the mixture component, dimension of the mean vector, and training speaker. In Equation (11), the s th row vector of Uspeaker, which is denoted as uspeaker;s, corresponds to the speaker weight of the s th speaker, thus the low-rank approximation of the s th speaker model is given by

$M s ≃ G K R × K D × K S × 3 u speaker ; s × 1 U mixture × 2 U dim + M ¯ .$
(12)

If we define the augmented speaker weight$W s K R × K D ≡ G K R × K D × K S × 3 u speaker ; s$, Equation (12) becomes

$M s ≃ W s × 1 U mixture × 2 U dim + M ¯ = U mixture W s U dim T + M ¯ .$
(13)

Thus, we express the model of a new speaker as

$M new = U mixture W new U dim T + M ¯ .$
(14)

For the given adaptation data O={o1,…,o T }, we derive the equation for finding the speaker weight in a maximum likelihood (ML) criterion:

$∑ t ∑ r γ r ( t ) C r − 1 U dim W new T ⏟ ≡ W new,aug T u mixture ; r T u mixture ; r = ∑ t ∑ r γ r ( t ) C r − 1 o t − m ¯ r T u mixture ; r$
(15)

where γ r (t) denotes the occupation probability of being at mixture r at t given O, C r the covariance matrix of the r th Gaussian component of an SI HMM (in this article, a diagonal covariance matrix is used); umixture;r and$m ¯ r$ denote the r th row vectors of Umixture and$M ¯$, respectively. In the above equation, Wnew, aug can be computed using a technique similar to MLLR adaptation and the weight of the new speaker is obtained by

$W ̂ new = W new , aug U dim$
(16)

which is plugged into Equation (14) to produce the model updated for the new speaker.

### 2.3 Probabilistic 2-mode analysis

The advantage of probabilistic 2-mode analysis over Tucker decomposition is similar to that of PPCA over standard PCA; probabilistic 2-mode analysis can deal with missing entries in the data tensor (although this is not the case in our experiments). In the modeling perspective, probabilistic 2-mode analysis assumes a distribution of latent variables, thus it is suitable for a MAP framework.

In this section, the ensemble of training models is expressed as

$M = { M s } s = 1 S .$
(17)

Assuming the HMM mean vectors of training speakers are drawn from the matrix-variate normal distribution, we derive the adaptation equation based on the probabilistic 2-mode analysis of training models. We use probabilistic 2-mode analysis, the second-order case of PTA, to decompose the training models expressed in matrix form. The latent tensor model is expressed as

$ℳ = W ∏ n = 1 N × n U n + ℳ mean + E$
(18)

where$W$ denotes the latent tensor, U n ’s the factor loading matrices,$ℳ mean$ the mean, and$E$ is the error/noise process. The 2-mode case of the latent tensor model is given by

$M = W × 1 U 1 × 2 U 2 + M mean + E$
(19)

which becomes, for the training models {M1,…,M S },

$M s = W s × 1 U 1 × 2 U 2 + M mean + E s = U mixture W s U dim T + M mean + E s$
(20)

where$W s ∈ ℝ K R × K D$ denotes the latent matrix,$U mixture ∈ ℝ R × K R$ and$U dim ∈ ℝ D × K D$ the factor loading matrices (K R R−1 and K D D−1), Mmean the mean, and E s the error/noise process. (Mode matrices and dimensions are defined as follows: U1=Umixture, U2=Udim, I1=R, I2=D, K1=K R , and K2=K D .) The distribution of W s is assumed to be a matrix-variate normal, i.e.,$W s ∼N( 0 K R × K D , I K R ⊗ I K D )$ where denotes the Kronecker product, and independent of E s whose elements follow$N(0, σ 2 )$. Figure4 shows the graphical model representing the probabilistic 2-mode model.

In Equation (20), it is computationally intractable to calculate U n ’s simultaneously. So, the following decoupled predictive density is defined:

$p M | M mean , { U n } n = 1 N , σ 2 ∝ ∏ n = 1 N p M × ̄ n U n T | t ̄ n , σ n 2$
(21)

where$t ̄ n ∈ ℝ I n × 1$ and$σ n 2$ denote the mean vector and noise variance, respectively, for mode n;$M × ̄ n U n T ≡M × 1 U 1 T … × n − 1 U n − 1 T × n + 1 U n + 1 T … × N U N T$, i.e., the product of M with all the mode matrices except mode n, which is called the contracted n-mode product. That is, the n th probabilistic function is defined as the projected M by all U j ’s expect U n . Given observed data M, the decoupled posterior probabilistic function is defined as

$p M mean , { U n } n = 1 N , σ 2 | M ∝ ∏ n = 1 N p t ̄ n , U n , σ n 2 , M , { U j } j = 1 , j ≠ n N .$
(22)

By Bayes’ theorem, the n th posterior distribution can be expressed in terms of the decoupled likelihood function and the decoupled prior distribution:

$p t ̄ n , U n , σ n 2 , M , { U j } j = 1 , j ≠ n N ∝ p M , { U j } j = 1 , j ≠ n N | t ̄ n , U n , σ n 2 p t ̄ n , U n , σ n 2 .$
(23)

Therefore, the decoupled predictive density is given by

$p M | M ∝ p M | M new , { U n } n = 1 N , σ 2 × p M mean , { U n } n = 1 N , σ 2 | M = ∏ n = 1 N p M × ̄ n U n T | t ̄ n , σ n 2 × p M , { U j } j = 1 , j ≠ n N | t ̄ n , U n , σ n 2 .$
(24)

$p t ̄ n , U n , σ n 2$ is dropped out for a fixed U n ). This is the 2-mode case of the PTA in. In our case, Equation (24) is given by

$p M | M ∝ p U dim T M T | t ̄ mixture , σ mixture 2 × p M , U dim | t ̄ mixture , U mixture , σ mixture 2 × p U mixture T M | t ̄ dim , σ dim 2 × p M , U mixture | t ̄ dim , U dim , σ dim 2 .$
(25)

Now, U n ’s are obtained by maximizing the following posterior distribution:

$p { U n } n = 1 N | M ≈ ∏ n = 1 N p U n | M × ̄ n U n T$
(26)

where$p U n |M × ̄ n U n T ≡ ∏ s = 1 S M s × ̄ n U n T$. The expectation-maximization (EM) algorithm is applied to compute U n ’s. The application of the EM algorithm to construct probabilistic 2-mode model is explained in the next section.

### 2.4 Construction of probabilistic 2-mode model for speaker adaptation

In Equation (20), for the given training models, the maximum likelihood (ML) estimate of Mmean is given as$M ¯ =(1/S) ∑ s M s$ and${ U n , σ n 2 }$ can be estimated as follows. First, let us define the followings: Let$t n ; j ∈ ℝ I n × 1$ be the j th column vector of

$T ( n ) = mat n M × ̄ n U n T$
(27)

for$1≤j≤ Ī n S$ ($Ī n = ∏ j = 1 , j ≠ n N I j$) and$x n ; j ∈ ℝ K n × 1$ be the j th column vector of

$X ( n ) = mat n M ∏ n = 1 N × n U n T .$
(28)

Let us suppose$t n | x n ∼N( U n x n + t ̄ n , σ n 2 I I n )$ and$x n ∼N( 0 K n × 1 , I K n )$. Then, by integrating out x n ,$t n ∼N( t ̄ n , G n )$ where$t ̄ n =1/( Ī n S) ∑ j = 1 Ī n S t n ; j$ and$G n = U n U n T + σ n 2 I I n$. Consequently,

$x n | t n ∼ N H n − 1 U n T ( t n − t ̄ n ) , σ n 2 H n − 1$
(29)

where$H n = U n T U n + σ n 2 I K n$. The right-hand side of Equation (26) becomes

$log p U n | M × ̄ n U n T ∝ − Ī n S 2 log | G n | + tr G n − 1 S n$
(30)

where$S n =1/( Ī n S−1) ∑ j = 1 Ī n S t n ; j − t ̄ n t n ; j − t ̄ n T$ and tr[ · ] denotes the trace of a matrix. Summing up for all the modes, we obtain the following log-likelihood function of the posterior distribution:

$L = ∑ n log p U n | M × ̄ n U n ∝ − ∑ n Ī n S 2 log | G n | + tr G n − 1 S n .$
(31)

The graphical model representation of the decoupled probabilistic model is shown in Figure5.

We seek to find U n ’s that maximize the log-likelihood function alternatingly. Mode matrices U1 and U2 are initialized with the results from the Tucker decomposition which minimizes the reconstruction error:

$Error = ∑ s M s − W s × 1 U 1 × 2 U 2 + M ¯ 2 .$
(32)

With the initial U1 and U2, the following procedure is performed for each mode (n=1,2).

Each training model is projected into mode matrices except mode n and expressed in a mode-n matrix:

$T s , ( n ) = mat n M s × ̄ n U n T .$
(33)

All the column vectors of$T s , ( n ) s = 1 S$ constitute the training data set:

${ t n ; j } , 1 ≤ j ≤ Ī n S.$
(34)

Then, with an initial estimate of$σ n 2$ (e.g., 0.005 was used in the experiments), the EM algorithm is iterated as follows until U n and$σ n 2$ converge.

E-step: From Equation (31), the expectation of the log-likelihood function of complete data w.r.t.$p x n ; j | t n ; j , t ̄ n , U n , σ n 2$ is given as

$〈 L c 〉 = ∑ n ∑ s E log p ( M s , W s | { U } j = 1 , j ≠ n N ) = ∑ n ∑ j = 1 I n S E log p ( t n ; j , x n ; j | { U } j = 1 , j ≠ n N )$
(35)

where

$log p ( t n ; j , x n ; j | { U } j = 1 , j ≠ n N ) = log p ( x n ; j ) + log p ( t n ; j | x n ; j ) ∝ − ∥ x n ; j ∥ 2 − I n 2 log ( σ n 2 ) − 1 σ n 2 ∥ t n ; j − U n x n ; j − t ̄ n ∥ 2 .$
(36)

So,

$〈 L c 〉 ∝ − ∑ n ∑ j = 1 Ī n S tr 〈 x n ; j x n ; j T 〉 + 1 σ n 2 ( t n ; j − t ̄ n ) T ( t n ; j − t ̄ n ) + I n 2 log ( σ n 2 ) + 1 σ n 2 tr U n T U n 〈 x n ; j x n ; j T 〉 − 2 σ n 2 〈 x n ; j 〉 T U n T ( t n ; j − t ̄ n )$
(37)

with the sufficient statistics are given as follows from Equation (29):

$〈 x n ; j 〉 = H n − 1 U n T t n ; j − t ̄ n 〈 x n ; j x n ; j T 〉 = σ n 2 H n − 1 + 〈 x n ; j 〉 〈 x n ; j 〉 T .$
(38)

M-step: Model parameters are updated by maximizing 〈Lc〉 w.r.t. U n and$σ n 2$. Setting$∂ U n 〈 L c 〉=0$ produces

$U n = ∑ j = 1 Ī n S t n ; j − t ̄ n 〈 x n ; j 〉 T ∑ j = 1 Ī n S 〈 x n ; j x n ; j T 〉 − 1 .$
(39)

Next, setting$∂ σ n 2 〈 L c 〉=0$ produces

$σ n 2 = 1 I n Ī n S ∑ j = 1 Ī n S ∥ t n ; j − t ̄ n ∥ 2 − 2 〈 x n ; j 〉 T U n T t n ; j − t ̄ n + tr 〈 x n ; j x n ; j T 〉 U n T U n .$
(40)

Essentially, the procedure applies PPCA to the data set {tn;j} for each mode.

### 2.5 Estimation of prior distribution

Given model parameters${ M ¯ , U n , σ n 2 }$, the weight matrix for the training speaker model M s is obtained by

$W s = M s − M ¯ ∏ n = 1 2 × n H n − 1 U n T = H 1 − 1 U mixture T M s − M ¯ H 2 − 1 U dim T T .$
(41)

From the set of weight matrices${ W s } s = 1 S$, the distribution of the weight is estimated. In deriving the adaptation equation in the MAP framework, the parameters for the prior distribution can be obtained in closed-form solutions if p(W) follows a conjugate distribution. Hence, we assume the prior distribution of the weight to be a matrix-variate normal:

$p ( W ) ∝ 1 | Σ | K D / 2 | Ψ | K R / 2 exp { − 1 2 tr [ ( W − W mean ) T Σ − 1 ( W − W mean ) Ψ − 1 ] } .$
(42)

Furthermore, the hyperparameters of p(W) can easily be estimated in an ML criterion if Ψ is known. So, Ψ is assumed to be the identity matrix, and the hyperparameters are estimated as:

$W ̂ mean = 1 S ∑ s W s = 0 K R × K D Σ ̂ = 1 S − 1 ∑ s W s W s T .$
(43)

### 2.6 Speaker adaptation in the MAP framework

Based on Equation (20), we express the model of a new speaker as

$M new = U mixture W new U dim T + M ¯ .$
(44)

For the given adaptation data O={o1,…,o T }, we estimate the adaptation parameter in a MAP criterion:

$Λ MAP = arg max Λ p ( Λ | O ) ∝ arg max Λ p ( O | Λ ) p ( Λ ) ∝ arg max Λ log p ( O | Λ ) + log p ( Λ )$
(45)

where Λ={Wnew} denotes the model parameter.

Using the EM algorithm, we obtain the following auxiliary Q-function to be optimized (discarding the terms that are independent of the model parameter):

$Q ( Λ , Λ ̂ ) = − 1 2 ∑ t ∑ r γ r ( t ) tr ( o t − m new ; r T ) T C r − 1 ( o t − m new ; r T ) − 1 2 tr ( W new U dim ) T Σ ̂ − 1 ( W new U dim )$
(46)

where Λ and$Λ ̂$ denote the current and updated model parameters, respectively, and$m new ; r = u mixture ; r W new U dim T + m ¯ r$. In finding the speaker weight, we compute Wnew, augWnewUdim, from which Wnew is obtained. Solving in this way, we can use the row-by-row technique in MLLR adaptation. Setting$∂ W new Q(Λ, Λ ̂ )=0$ yields the following equation:

$∑ t ∑ r γ r ( t ) C r − 1 U dim W new T ⏟ ≡ W new, aug T u mixture ; r T u mixture ; r + U dim W new T ⏟ ≡ W new, aug T Σ ̂ − 1 = ∑ t ∑ r γ r ( t ) C r − 1 o t − m ¯ r T u mixture ; r .$
(47)

The above equation can be solved for Wnew, aug in a similar way to MLLR adaptation in: we define the followings:

$V r = ∑ t γ r ( t ) C r − 1 D r = u mixture ; r T u mixture ; r G ( i ) = ∑ r v r ( i , i ) D r Z = ∑ t ∑ r γ r ( t ) C r − 1 o t − m ¯ r T u mixture ; r Σ ( i ) = 1 S − 1 ∑ s w s ; i w s ; i T$
(48)

where v r (i,i) denotes the (i,i) element of V r and ws;i the i th column vector of Ws, augW s Udim. Then, the speaker weight can be computed:

$w new, aug , ( i ) T = G ( i ) + Σ ( i ) − 1 − 1 z ( i ) T , i = 1 , … , D$
(49)

where wnew, aug, (i) denotes the i th row of Wnew, aug and z(i) the i th row vector of Z. The method becomes similar to MAPLR adaptation in. Finally, the speaker weight is obtained as

$W ̂ new = W new, aug U dim +$
(50)

where [ · ]+ denotes the pseudoinverse of a matrix. The weight is plugged into Equation (44) to produce the model adapted to the new speaker.

### 2.7 Speaker adaptation techniques compared in the experiments

In this section, we briefly review the speaker adaptation techniques compared with the probabilistic 2-mode analysis based method: eigenvoice adaptation, MLLR adaptation, and MAPLR adaptation.

In eigenvoice adaptation, the collection of HMM mean vectors of speaker s is arranged in an (R D)×1 vector:

$μ s = μ s ; 1 μ s ; 2 ⋮ μ s ; R .$
(51)

Then, the set of S supervectors, {μ1,…,μ S }, is decomposed by PCA to produce the adaptation model

$μ new = Φ w new + μ ̄$
(52)

where Φ=[ϕ1ϕ K ], the basis matrix consisting of the K dominant eigenvectors from PCA, and$μ ̄ =1/S ∑ s μ s$. The K×1 weight vector can be obtained by maximizing the likelihood of the adaptation data, which is given by

$w ̂ new = ∑ t ∑ r γ r ( t ) Φ r T C r − 1 Φ r − 1 × [ ∑ t ∑ r γ r ( t ) Φ r T C r − 1 o t − μ ̄ r ]$
(53)

where Φ r and$μ ̄ r$ denote the D×K submatrix and D×1 subvector corresponding to the r th mixture of Φ and$μ ̄$, respectively.

In MLLR adaptation, the updated model for a new speaker is obtained by linearly transforming the SI model (assuming a global regression matrix):

$μ new , r = W new ξ r , ξ r = ω μ SI, r$
(54)

where μSI, r denotes the mean vector of the SI HMM corresponding to mixture r and ω is the bias offset term: ω=1 to include the term and ω=0 otherwise (ω=1 in our experiments). The D×(D+1) transformation matrix can be obtained in an ML criterion, which yields the following equation:

$∑ t ∑ r γ r ( t ) C r − 1 o t ξ r T = ∑ t ∑ r γ r ( t ) C r − 1 W new ξ r ξ r T .$
(55)

The above equation can be solved for Wnew:

$w ̂ new, ( i ) T = G ( i ) − 1 z ( i ) T , i = 1 , … , D$
(56)

where$w ̂ new, ( i )$ and z(i) denote the i th row vectors of$W ̂ new$ and Z, respectively; G(i) and Z are defined as:

$V r = ∑ t γ r ( t ) C r − 1 D r = ξ r ξ r T G ( i ) = ∑ r v r ( i , i ) D r Z = ∑ t ∑ r γ r ( t ) C r − 1 o t ξ r T$
(57)

where v r (i,i) denotes the (i,i) element of V r .

In MAPLR adaptation, the prior for the transformation matrix is used in the MLLR framework. The parameters for the prior are obtained from the MLLR transformation matrices of training speakers {W1,…,W S }:

$w ̄ ( i ) = 1 S ∑ s w s , ( i ) Σ ( i ) = 1 S − 1 ∑ s w s , ( i ) − w ̄ ( i ) T w s , ( i ) − w ̄ ( i )$
(58)

where ws, (i) denotes the i th row vector of W s . Then, the transformation matrix for a new speaker is obtained in a MAP criterion. Deriving the equation in the same way as above, we can obtain the following:

$w ̂ new, ( i ) T = G ( i ) + Σ ( i ) − 1 − 1 z ( i ) + w ̄ ( i ) Σ ( i ) − 1 T .$
(59)

## 3 Experiments

We carried out the large-vocabulary continuous-speech recognition (LVCSR) experiments using the Wall Street Journal corpus WSJ0. In building the SI model, we used 12754 utterances of 101 speakers from the corpus. As the acoustic feature vector, we used the 39-dimensional vector consisting of 13-dimensional mel-frequency cepstral coefficients (MFCCs) including the 0th cepstral coefficient, their derivative coefficients, and their acceleration coefficients. The feature vector was extracted with the 20-ms Hamming window with the frame sliding of 10 ms. Using the HMM toolkit (HTK), we built a tied-state triphone model (word-internal triphones) with 3472 tied states and 8-mixture Gaussian.

To build training models for constructing bases, we transformed the SI model by MLLR adaptation using 32 regression classes followed by maximum a posteriori (MAP) adaptation. We used the 101 adapted models to build the Tucker decomposition and probabilistic tensor based models as well as eigenvoice.

For adaptation and recognition test, we used Nov’92 5K non-verbalized adaptation and test sets. The number of testing speakers was 8; adaptation set was used for adaptation and testing set of 330 sentences was used for recognition test (the number of testing utterances per speaker was about 40). The length of an adaptation sentence was about 6 s and the adaptation was performed in supervised mode. In recognition test, we used WSJ 5K non-verbalized 5k closed-vocabulary set and WSJ standard 5K non-verbalized closed bigram.

We think that the performance improvement of the proposed method over MLLR or MAPLR adaptation comes from the use of basis vectors and speaker weight of large dimension. Additionally, we think that the performance improvement of the probabilistic 2-mode based method in the MAP framework over the Tucker decomposition based method in the ML framework for small amounts of adaptation data (e.g., 1 adaptation sentence) is due to its constraint on the weight. If the amount of adaptation data is small (e.g., 1 adaptation sentence), the weight cannot be reliably estimated in the ML framework where the weight is estimated using only adaptation data without constraint, as done in the Tucker decomposition based method. The results confirm that constraint on the weight in the MAP framework can produce better model when the amount of adaptation data is small.

The selection of appropriate dimensions of model parameters (e.g., K R and K D ) in the probabilistic 2-mode analysis depends on the training models and also available adaptation data. The selection of model parameters affects the performance of the system, but how to choose the optimum model parameters is not obvious, which needs a further study.

## 4 Conclusions

In this article, we applied probabilistic tensor analysis to the adaptation of HMM mean vectors to a new speaker. The training models consisted of the mean vectors of HMMs expressed in matrix form and the training set was decomposed by probabilistic 2-mode analysis. The prior distribution of the adaptation parameter was estimated from the training models. Then, the speaker adaptation equation was derived in the MAP framework. Compared with the speaker adaptation method based on Tucker 3-mode decomposition in the ML framework, the proposed method further improved the performance for small amounts of adaptation data.

## Abbreviations

ALS:

Alternating Least Squares

ASR:

Automatic Speech Recognition

EM:

Expectation-Maximization

HMM:

Hidden Markov Model

HTK:

HMM Toolkit

LVCSR:

Large-Vocabulary Continuous-Speech Recognition

MAP:

Maximum A Posteriori

MAPLR:

Maximum A Posteriori Linear Regression

MFCC:

Mel-Frequency Cepstral Coefficient

ML:

Maximum Likelihood

MLLR:

Maximum Likelihood Linear Regression

PCA:

Principal Component Analysis

PPCA:

Probabilistic Principal Component Analysis

PTA:

Probabilistic Tensor Analysis

SD:

Speaker-Dependent

SI:

Speaker-Independent

SVD:

Singular Value Decomposition

## References

1. 1.

Rabiner LR: A tutorial on hidden Markov models and selected applications in speech recognition. Proc. IEEE 1989, 77(2):257-286. 10.1109/5.18626

2. 2.

Gales M, Young S: The application of hidden Markov models in speech recognition. Found. Trends Signal, Process 2008, 1(3):195-304.

3. 3.

Kuhn R, Junqua J-C, Nguyen P, Niedzielski N: Rapid speaker adaptation in eigenvoice space. IEEE Trans. Speech Audio Process 2000, 8(6):695-707. 10.1109/89.876308

4. 4.

Kolda TG, Bader BW: Tensor decompositions and applications. SIAM Rev 2009, 51(3):455-500. 10.1137/07070111X

5. 5.

Jeong Y: Speaker adaptation based on the multilinear decomposition of training speaker models. In Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing. TX: Dallas; 2010:4870-4873.

6. 6.

Leggetter CJ, Woodland PC: Maximum likelihood linear regression for speaker adaptation of continuous density hidden Markov models. Comput. Speech Lang 1995, 9(2):171-185. 10.1006/csla.1995.0010

7. 7.

Jeong Y: Acoustic model adaptation based on tensor analysis of training models. IEEE Signal Process. Lett 2011, 18(6):347-350.

8. 8.

Tao D, Song M, Li X, Shen J, Sun J, Wu X, Faloutsos C, Maybank SJ: Bayesian tensor approach for 3-D face modeling. IEEE Trans. Circ. Syst. Video Technol 2008, 18(10):1397-1410.

9. 9.

Tipping ME, Bishop CM: Probabilistic principal component analysis. J. R. Stat. Soc. Ser. B-Stat. Methodol 1999, 61(3):611-622. 10.1111/1467-9868.00196

10. 10.

Gauvain J-L, Lee C-H: Maximum a posteriori estimation for multivariate Gaussian mixture observations of Markov chains. IEEE Trans. Speech Audio Process 1994, 2(2):291-298. 10.1109/89.279278

11. 11.

Chesta C, Siohan O, Lee C-H: Maximum a posteriori linear regression for hidden Markov model adaptation. In Proceedings of EUROSPEECH. Hungary: Budapest; 1999:211-214.

12. 12.

Carroll JD, Chang JJ: Analysis of individual differences in multidimensional scaling via an N-way generalization of “Eckart-Young” decomposition. Psychometrika 1970, 35(3):283-319. 10.1007/BF02310791

13. 13.

Gupta AK, Nagar DK: Matrix Variate Distributions. Boca Raton, FL: Chapman and Hall/CRC; 1999.

14. 14.

Bader BW, Kolda TG: Algorithm 862: MATLAB tensor classes for fast algorithm prototyping. ACM Trans. Math. Softw 2006, 32(4):635-653. 10.1145/1186785.1186794

15. 15.

Dempster AP, Laird NM, Rubin DB: Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. Ser. B-Stat. Methodol 1977, 39(1):1-38.

16. 16.

Gupta AK, Varga T: Elliptically Contoured Models in Statistics. Norwell, MA: Kluwer; 1993.

17. 17.

Siohan O, Chesta C, Lee C-H: Joint maximum a posteriori adaptation of transformation and HMM parameters. IEEE Trans. Speech Audio Process 2001, 9(14):417-428.

18. 18.

Paul DB, Baker JM: The design for the Wall Street Journal-based CSR corpus. In Proceedings of DARPA Speech and Natural Language Workshop. TX: Austin; 1992:357-362.

19. 19.

Young S, Evermann G, Kershaw D, Moore G, Odell J, Ollason D, Povey D, Valtchev V, Woodland P: The HTK Book, Version 3.2. England: Cambridge University Engineering Department; 2002.

## Author information

Authors

### Corresponding author

Correspondence to Yongwon Jeong.

### Competing interests

The author declares that they have no competing interests.

## Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

## Rights and permissions

Reprints and Permissions

Jeong, Y. Speaker adaptation in the maximum a posteriori framework based on the probabilistic 2-mode analysis of training models. J AUDIO SPEECH MUSIC PROC. 2013, 7 (2013). https://doi.org/10.1186/1687-4722-2013-7 