Skip to main content

Context-dependent acoustic modeling based on hidden maximum entropy model for statistical parametric speech synthesis


Decision tree-clustered context-dependent hidden semi-Markov models (HSMMs) are typically used in statistical parametric speech synthesis to represent probability densities of acoustic features given contextual factors. This paper addresses three major limitations of this decision tree-based structure: (i) The decision tree structure lacks adequate context generalization. (ii) It is unable to express complex context dependencies. (iii) Parameters generated from this structure represent sudden transitions between adjacent states. In order to alleviate the above limitations, many former papers applied multiple decision trees with an additive assumption over those trees. Similarly, the current study uses multiple decision trees as well, but instead of the additive assumption, it is proposed to train the smoothest distribution by maximizing entropy measure. Obviously, increasing the smoothness of the distribution improves the context generalization. The proposed model, named hidden maximum entropy model (HMEM), estimates a distribution that maximizes entropy subject to multiple moment-based constraints. Due to the simultaneous use of multiple decision trees and maximum entropy measure, the three aforementioned issues are considerably alleviated. Relying on HMEM, a novel speech synthesis system has been developed with maximum likelihood (ML) parameter re-estimation as well as maximum output probability parameter generation. Additionally, an effective and fast algorithm that builds multiple decision trees in parallel is devised. Two sets of experiments have been conducted to evaluate the performance of the proposed system. In the first set of experiments, HMEM with some heuristic context clusters is implemented. This system outperformed the decision tree structure in small training databases (i.e., 50, 100, and 200 sentences). In the second set of experiments, the HMEM performance with four parallel decision trees is investigated using both subjective and objective tests. All evaluation results of the second experiment confirm significant improvement of the proposed system over the conventional HSMM.

1 Introduction

Statistical parametric speech synthesis (SPSS) has dominated speech synthesis research area over the last decade [1, 2]. It is mainly due to SPSS advantages over traditional concatenative speech synthesis approaches; these advantages include the flexibility to change voice characteristics [35], multilingual support [68], coverage of acoustic space [1], small footprint [1], and robustness [4, 9]. All of the above advantages stem from the fact that SPSS provides a statistical model for acoustic features instead of using original speech waveforms. However, these advantages are achieved at the expense of one major disadvantage, i.e., degradation in the quality of synthetic speech [1]. This shortcoming results from three important factors: vocoding distortion [1013], accuracy of statistical models [1425], and accuracy of parameter generation algorithms [2628]. This paper is an attempt to alleviate the second factor and improve the accuracy of statistical models. Most of the researches carried out to improve the acoustic modeling performance aimed to develop systems that generate natural and high-quality speech using large training speech databases (more than 30 min) [18, 21, 22]. Nevertheless, there exist a great number of under-resourced languages (such as Persian) for which only limited amount of data are available. To alleviate this shortcoming, we target developing a statistical approach that leads to an appropriate speech synthesis system not only with large but also with small training databases.

Every SPSS system consists of two distinct phases, namely training and synthesis [1, 2]. In the training phase, first acoustic and contextual factors are extracted for the whole training database using a vocoder [12, 29, 30] and a natural language pre-processor. Next, the relationship between acoustic and contextual factors is modeled using a context-dependent statistical approach [1425]. Synthesis phase starts with a parameter generation algorithm [2628] that exploits trained context-dependent statistical models and aims to generate realistic acoustic feature trajectories for a given input text. Acoustic trajectories are then fed into the same vocoder used during the training phase in order to generate the desired synthesized speech.

In the most predominant statistical parametric approach, spectrum, excitation, and duration of speech are expressed concurrently in a unified framework of context-dependent multi-space probability distribution hidden semi-Markov model (HSMM)[14]. More specifically, a multi-space probability distribution [17] is estimated for each leaf node of decision trees [31]. These decision tree-based structures split contextual space into a number of non-overlapped clusters which form multiple groups of context-dependent HMM states, and each group shares the same output probability distribution [31]. In order to capture acoustic variations accurately, the model has to be able to express a large number of robust distributions [19, 20]. Decision trees are not efficient for such expression because increasing the number of distributions by growing the tree reduces the population of each leaf and consequently reduces the robustness of the distributions. This problem stemmed from the fact that decision tree assigns each HMM state to an only one cluster (small region in contextual space), therefore, each state contributes in modeling just one distribution. In other words, the decision tree structure makes the models match training data just in non-overlapped regions which are expressed through decision tree terminal nodes [31]. In the case of limited training data, the decision tree would be small, so it cannot split contextual factor space sufficiently. In this case, the accordance between model and data is not sufficient, and therefore, the speech synthesis system generates unsatisfactory output. Accordingly, it is clear that by extending the decision tree in such a way that each state affects multiple distributions (larger portion of the contextual space), the generalization to unseen models will be improved. The main idea of this study is to extend non-overlapped regions of one decision tree to overlapped regions of multiple decision trees and hence exploit contextual factors more efficiently.

A large number of research works have already been performed to improve the quality of basic decision tree-clustered HSMM. Some of them are based on a model adaptation technique. This latter method exploits an invaluable prior knowledge attained from an average voice model [3], and adapts this general model using an adaptation algorithm such as maximum likelihood linear regression (MLLR)[32], maximum a posteriori (MAP)[33], and cluster adaptive training (CAT)[21]. However, working with average voice models is difficult for under-resourced languages since building such general model needs remarkable efforts to design, record, and transcribe a thorough multi-speaker speech database [3]. To alleviate the data sparsity problem in under-resourced languages, speaker and language factorization (SLF) technique can be used [34]. SLF attempts to factorize speaker-specific and language-specific characteristics in training data and then model them using different transforms. By representing the speaker attributes by one transform and language characteristics by a different transform, the speech synthesis system will be able to alter language and speaker separately. In this framework, it is possible to exploit the data from different languages to predict speaker-specific characteristics of the target speaker, and consequently, the data sparsity problem will be alleviated. Authors in [15, 16] also developed a new technique by replacing maximum likelihood (ML) point estimate of HSMM with a variational Bayesian method. Their system was shown to outperform HSMM when the amount of training data is small. Other notable structures used to improve statistical modeling accuracy are deep neural networks (DNNs)[18]. The decision tree structure is not efficient enough to model complicated context dependencies such as XORs or multiplexers [18]. To model such complex contextual functions, the decision tree has to be excessively large, but DNNs are capable to model complex contextual factors by employing multiple hidden layers. Additionally, a great number of overlapped contextual factors can be fed into a DNN to approximate output acoustic features, so DNNs are able to provide efficient context generalization. Speech synthesis based on Gaussian process regression (GPR) [35] is another novel approach that has recently been proposed to overcome HMM-based speech synthesis limitations. The GPR model predicts frame-level acoustic trajectories from frame-level contextual factors. The frame-level contextual factors include the relative position of the current frame within the phone and some articulatory information. These frame-level contextual factors are employed as the explanatory variable in GPR. The frame-level modeling of GPR removes the inaccurate stationarity assumption of state output distribution in HMM-based speech synthesis. Also, GPR can directly represent the complex context dependencies without using parameter tying by decision tree clustering; therefore, it is capable of improving context generalization.

Acoustic modeling with contextual additive structure has also been proposed to represent dependencies between contextual factors and acoustic features more precisely [19, 20, 23, 32, 3640]. In this structure, acoustic trajectories are considered to be a sum of independent acoustic components which have different context dependencies (different decision trees have to be trained for those components). Since the mean vectors and covariance matrices of the distribution are equal to the sum of mean vectors and covariance matrices of additive components, the model would be able to exploit contextual factors more efficiently. Furthermore, in this structure, each training data sample contributes to modeling multiple mean vectors and covariance matrices. Many papers applied the additive structure just for F0 modeling [3740]. Authors in [37] proposed an additive structure with multiple decision trees for mean vectors and a single tree for variance terms. In this paper, for different additive components, different sets of contextual factors were used and multiple trees were built simultaneously. In [40], multiple additive decision trees are also employed, but they train this structure using minimum generation error (MGE) criterion. Sakai [38] defines an additive model with three distinct layers, namely intonational phrase, word-level, and pitch-accent layers. All of these components were trained simultaneously using a regularized least square error criterion. Qian et al. [39] propose to use multiple additive regression trees with a gradient-based tree-boosting algorithm. Decision trees are trained in successive stages to minimize the error squares. Takaki et al. [19, 20] applied additive structure for spectral modeling and reported that the computational complexity of this structure is extremely high for full context labels as used in speech synthesis. To alleviate this issue, they proposed two approaches: covariance parameter tying and a likelihood calculation algorithm using matrix inversion lemma [19]. Despite all the advantages, this additive structure may not match training data accurately because once training is done, the first and second moments of the training data and model may not be exactly the same in some regions.

Another important problem of conventional decision tree-clustered acoustic modeling is difficulty in capturing the effect of weak contextual factors such as word-level emphasis [23, 36]. It is mainly because weak contexts have less influence on the likelihood measure [23]. One clear approach to address this issue is to construct the decision tree in two successive steps [36]. In the first step, all selections are done among weak contextual factors, and in the second step, the remaining questions are adopted [36]. This procedure can effectively exploit weak contextual factors, but it leads to a reduction in the amount of training data available for normal contextual factors. Context adaptive training with factorized decision trees [23] is another approach that can exploit weak context questions efficiently. In this system, a canonical model is trained using normal contextual factors and then a set of transforms is built by weak contextual factors. In fact, canonical models and transforms, respectively, represent the effects of normal and weak contextual factors [23]. However, this structure also improves context generalization of conventional HMM-based synthesis by exploiting adaptation techniques.

This paper introduces a maximum entropy model (MEM)-based speech synthesis. MEM [41] has been demonstrated to be positively effective in numerous applications of speech and natural language processing such as speech recognition [42], prosody labeling [43], and part-of-speech tagging [44]. Accordingly, the overall idea of this research is to improve HSMM context generalization by taking advantage of a distribution which not only matches training data in many overlapped contextual regions but also is optimum in the sense of an entropy criterion. This system has the potential to model the dependencies between contextual factors and acoustic features such that each training sample contributes to train multiple sets of model parameters. As a result, context-dependent acoustic modeling based on MEM could lead to a promising synthesis system even for limited training data.

The rest of the paper is organized as follows. Section 2 presents HSMM-based speech synthesis. The hidden maximum entropy model (HMEM) structure and the proposed HMEM-based speech synthesis system are explained in Section 3. Section 4 is dedicated to experimental results. Finally, Section 5 concludes this paper.

2 HSMM-based speech synthesis

This section aims to explain the predominant statistical modeling approach applied in speech synthesis, i.e., context-dependent multi-space probability distribution left-to-right without skip transitions HSMM[3, 14] (simply called HSMM in the remainder of this paper). The discussion presented in this section provides a preliminary framework which will be used as a basis to introduce the proposed HMEM technique in Section 3. The most significant drawback of HSMM, namely inadequate context generalization, is also pointed out.

2.1 HSMM structure

HSMM is a hidden Markov model (HMM) having explicit state duration distribution instead of self-state transition probabilities. Figure 1 illustrates the standard HSMM. As it can be observed, HSMM initially partitions acoustic parameter (observation) trajectories into a fixed number of time slices (so-called states) in order to moderate the undesirable influence of non-stationarity. Note that state durations are latent variables and have to be trained in an unsupervised manner. An N-state HSMM λ is specified by a set of state output probability distributions b i i = 1 N and a complementary set of state duration probability distributions p i i = 1 N . To model these distributions, a number of distinct decision trees are used for output and duration probability distributions. Conventionally, different trees are trained for different states [31]. These trees cluster the whole contextual factor space into a large number of tiny regions which are expressed by terminal nodes. Thereafter, in each terminal node, the output distribution b i (∙) is modeled by a multi-space probability distribution, and similarly, a typical Gaussian distribution is considered for the duration probability p i (∙) [14].

Figure 1

HSMM structure and its output and duration probability distribution.

To handle the absence of fundamental frequency in unvoiced regions, multi-space probability distribution (MSD) is used for output probability distribution [17]. In accordance with commonly used synthesizers, this paper assumes that acoustic sample space consists of G spaces. Each of these spaces, specified by an index g, represents an n g dimensional real space, i.e., n g . Each observation vector o t has a probability w g to be generated by the g th space iff the dimensionality of o t is identical to n g . In other words, we have

b i o t = g ϵ S o t w ig b i | g o t , b i | g o t = N n g o t ; μ ig , Σ ig ,
p i d = N 1 d ; m i , σ i 2 ,

where S(o t ) represents a set of all space indexes with the same dimensionality of o t , and where N l . ; μ , Σ denotes an l-dimensional Gaussian distribution with mean μ, and covariance matrix ∑ ( N 0 is defined to be 1). Furthermore, the output probability distribution of the i th state and g th space is denoted by b i|g (o t ) which is a Gaussian distribution with mean vector μ ig and covariance matrix ∑ ig . Also, m i and σ i 2 represent mean and variance of the state duration probability.

Regarding the method for providing context dependency, it should be noted that HSMM normally offers binary decision trees and acoustic models are established for each leaf of these trees, separately [45, 46]. Suppose f and L are contextual functions based on a decision tree Y and are defined as

f l i ; Υ def ¯ ¯ 1 if context of the i th state l th leaf of Υ 0 if context of the i th state l th leaf of Υ , L Υ def ¯ ¯ number of leaves in Υ

Applying the above functions, all model parameters of Equations 1 and 2 can be expressed by linear combinations of model parameters defined for each terminal node. More precisely,

Σ ig = l = 1 L Υ o f l i ; Υ o Σ ig l , μ ig = l = 1 L Υ o f l i ; Υ o μ ig l , w ig = l = 1 L Υ o f l i ; Υ o w ig l , σ i = l = 1 L Υ o f l i ; Υ d σ i l , m i = l = 1 L Υ o f l i ; Υ d m i l ,

where Yo and Yd are decision trees trained for modeling output observation vectors and state durations. All symbols with superscript l indicate model parameters defined for the l th leaf.

2.2 HSMM likelihood

Having described the HSMM structure, we can now probe the exact expression for model likelihood or the probability of the observation sequence O = [o1,o2,…,o T ] as [14]:

P O | λ = i = 1 N j = 1 , j i N d = 1 t α t d j p i d s = t d + 1 t b i o s β t i

where this equality is valid for every value of [1,T]. Also, α t (i) and βt(i) are partial forward and backward probability variables that are calculated successively from their previous or next values as follows [3, 14]:

α t i = d = 1 t j = 1 , j i N α t d j p i d s = t d + 1 t b i o s ,
β t i = d = 1 T t j = 1 , j i N p j d s = t + 1 t + d b j o s β t + d j ,

where the initial forward and backward variables for every state indexes i are α0(i)-1 and β T (i) = 1.

2.3 HSMM parameter re-estimation

The ML criterion is commonly used to estimate model parameters of HSMM. However, we are not aware of latent variables, i.e., state durations and space indexes; therefore, an expectation maximization (EM) algorithm has to be adopted. Applying EM algorithm leads to the following re-estimation formulas [14]:

μ ^ ig l = k f l i ; Υ o t = 1 T γ t i , g o t k f l i ; Υ o t = 1 T γ t i , g , ^ ig l = k f l i ; Υ o t = 1 T γ t i , g o t μ ^ ig l o t μ ^ ig l T k f l i ; Υ o t = 1 T γ t i , g , w ^ ig l = k f l i ; Υ o t = 1 T γ t i , g h = 1 G k f l i ; Υ o t = 1 T γ t i , h , m ^ i l = k f l i ; Υ d t = 1 T d = 1 t χ t d i d k f l i ; Υ d t = 1 T d = 1 t χ t d i , σ ^ i l 2 = k f l i ; Υ d t = 1 T d = 1 t χ t d i d m ^ i l 2 k f l i ; Υ d t = 1 T d = 1 t χ t d i ,

where γ t (i,g) denotes the posterior probability of being in state i and space g at time t, and χ t d i is the probability of occupying the i th state from time td + 1 to t. The following equations calculate the above probabilities:

γ t i , g = 1 P o λ t 0 = 1 t 1 t 1 = t T j = 1 , j i N α t d j p i d b i g o t s = t 0 , s t t 1 b i o s β t 1 i , x t d i = 1 p o λ j = 1 , j i N α t d j p i d s = t d + 1 t b i o s β t i .

2.4 Inefficient context generalization

A major drawback of decision tree-clustered HSMM can now be clarified. Suppose we have only two real contextual factors, f1 and f2. Figure 2 shows a sample decision tree and the regions represented by its terminal nodes. By training HSMM, the model matches training data in all non-overlapped regions expressed by the terminal nodes. However, there is no guarantee that this accordance is held for overlapped regions such as the region R in Figure 2.

Figure 2

Sample decision tree and regions represented by its terminal nodes. (A) An example of decision tree with just three questions. (B) Regions that are classified by the tree and an arbitrary region (named R).

It can be noticed from the definition of function fi(c;Y) in Equation 3 that this function can be viewed as a set of L(Y) non-overlapped binary contextual factors. The fact that these contextual factors are non-overlapped leads to the insufficient context generalization, because this fact makes each training sample contribute to the model of only one leaf and only one Gaussian distribution. Hence, by extending fi(c;Y) to overlapped contextual factors, more efficient context generalization capabilities could be achieved. Section 3 proposes an approach which enables the conventional structure to model the overlapped contextual factors and thus improves the modeling performance of unseen contexts.

3. Hidden maximum entropy model

The goal of this section is to develop a context-dependent statistical model for acoustic parameters with adequate context generalization. The previous section on HSMM revealed that inappropriate generalization stemmed from the application of non-overlapped features only. Consequently, relating acoustic parameters to contextual information by incorporating overlapped features could improve generalization efficiency. This section proposes HMEM to establish this relation.

3.1 HMEM structure

The proposed HMEM technique exploits exactly the same structure and graphical model as the original HSMM, and thus, the model likelihood expression given by Equation 5 is also valid for HMEM. The only difference between HSMM and HMEM is the way they incorporate contextual factors in output and duration probability distributions (i.e., b i i = 1 N , p i i = 1 N ). HSMM builds a decision tree and then trains a Gaussian distribution for each leaf of the tree. On the contrary, HMEM obeys the maximum entropy modeling approach which will be described in the next subsection.

3.1.1 Maximum entropy modeling

Let us now derive a simple maximum entropy model. Suppose an ℓ-dimensional random vector process with output x that may be influenced by some contextual information c. Our target is to construct a stochastic model that precisely predicts the behavior of x, when c is given, i.e., P(x|c). Maximum entropy principle first imposes a set of constraints on P(x|c) and then chooses a distribution as close as possible to a uniform distribution by maximizing the entropy criterion [41]. In fact, this method will find the least biased distribution among all distributions that satisfy our constraints. In other words,

P ^ x | c def ¯ ¯ argmax P P subject to a set of constraint ,

employed constraints make the model preserve some context-dependent statistics of the training data. (P) represents entropy criterion [41] that is calculated as

P def ¯ ¯ x all possible c P x , c log P x , c dx .

Computing the above expression is extremely complex because there are a large number of contextual factors and all possible values of c are not calculable. However, authors in [41] applied the following approximation for P(x, c):

P x , c = P ˜ c P x | c .

where P ˜ c denotes empirical probability which can be calculated directly using the training database [41]. The above approximation simplifies the entropy expression as

P = x all c in database P ˜ c P x | c log P x | c dx all c in database P ˜ c log P ˜ c ,

where the second term is constant and does not affect the optimization problem. Therefore, we have

P = all c in database P ˜ c x P x | c log P x | c dx .

Additionally, we adopt a set of L f predefined binary contextual factors, f l (c), and another set of L g binary contextual factors, g l (c), that both of them may be highly overlapped. In order to obtain a Gaussian distribution for P ^ x | c and extend the conventional HSMM distribution, first- and second-order context-dependent moments expressed in Equation 14 are considered for the constraints.

P ^ x | c def ¯ ¯ argmax P P

subject to following constraints:

1 l L f E f l c x = E ˜ f l c x 1 l L g E g l c x x T = E ˜ g l c x x T For all possible c x P x | c dx = 1 ,

where E and E ^ indicate real and empirical mathematical expectations given in the following equations:

E ˜ f l c x = all c in database P ˜ c f l c x c ,
E f l c x = all c in database P ˜ c f l c x xP x , c dx

where x(c) denotes the realization of ℓ-dimensional random vector x for the context c in the database. If there are multiple realizations for x, x(c) will be obtained by taking the average over those values. In sum, the proposed context-dependent acoustic modeling approach obtains the smoothest (maximum entropy) distribution that captures first-order moments of training data in L f regions indicated by f l c l = 1 L f and second-order moments of data computed in g l c l = 1 L g .

In order to solve the optimization problem expressed by Equation 10, the Lagrange multipliers method is applied. This method defines a new optimization function as follows:

P ^ x | c = argmax P P + l = 1 L f u l T E f l c x E ˜ f l c x + l = 1 L g E g l c x T H l x E ˜ g l c x T H l x ,

where u l denotes a vector of Lagrange multipliers for satisfying the l th first-order moment constraints and H l is a matrix of Lagrange multipliers for satisfying the l th second-order moment constraints. Taking derivatives of the above function with respect to P leads to the following equality.

all c P ˜ c x log P x | c + u T x + x T Hx + const . dx = 0 H def ¯ ¯ l = 1 L g g l c H l , u def ¯ ¯ l = 1 L f f l c u l .

Therefore, one possible solution that maximizes entropy with the constraint of Equation 15 using Lagrange multipliers can be expressed as:

P ^ x | c = 1 det 2 π H 1 0.5 exp x 1 2 x + 1 2 H 1 u T H x + 1 2 H 1 u , H def ¯ ¯ l = 1 L g g l c H l , u def ¯ ¯ l = 1 L f f l c u l ,

where H l and u l are model parameters related to the l th contextual factors g l (c) and f l (c), respectively. H l is an ℓ-by-ℓ matrix and u l is an ℓ-dimensional vector. When f l (c) becomes 1 (i.e., it is active), u l affects the distribution; otherwise, it has no effect on the distribution. In fact, Equation 19 is nothing but the well-known Gaussian distribution with mean vector –0.5H-1u, and covariance matrix H-1, both calculated from a specific context-dependent combination of model parameters. Indeed, the main difference of MEM in comparison with other methods such as spectral additive structure [19, 20] is that mean and variance in MEM are not a linear combination of other parameters. This type of combination enables MEM to match training data in all overlapped regions.

This form of context-dependent Gaussian distribution presents a promising flexibility in utilizing contextual information. On one hand, using detailed and non-overlapped contextual factors such as features defined by Equation 3 (decision tree terminal node indicators) generates context-dependent Gaussian distributions which are identical to those used in conventional HSMM. These distributions have straightforward and efficient training procedure but suffer from insufficient context generalization capabilities. On the other hand, incorporating general and highly overlapped contextual factors overcomes the latter shortcoming and provides efficient context generalization, but its training procedure becomes more computationally complex. In the case of highly overlapped contextual factors, an arbitrary context activates several contextual factors, and hence, each observation vector is involved in modeling several model parameters.

3.1.2 ME-based modeling vs. additive modeling

At first glance, the contextual additive structure [19, 20, 32, 37] seems to have the same capabilities as the proposed ME-based context-dependent acoustic modeling. Therefore, to clarify their differences, this section compares HMEM with the additive structure through a very simple example.

In this example, the goal is to model a one-dimensional observation value using both ME-based modeling and a contextual additive structure. Due to the prime importance of mean parameters in HMM-based speech synthesis [47], we investigate the difference between mean values predicted by two systems.

Figure 3A shows a three-dimensional contextual factor space (c1-c2-c3) which is clustered by an additive structure. The additive structure consists of three different additive components with three different decision trees, namely Q1, Q2, and Q3. Each tree has a simple structure with just one binary question that splits a specific dimension of the contextual factor space into two regions. Each region is represented by a leaf node, and inside that leaf node, a mean parameter of each additive component is written. As it is depicted in the figure, these trees split contextual factor space into eight different cubic clusters. Mean values estimated for these cubic clusters are computed by adding mean values of additive components.

Figure 3

Contextual factor space clustered by (A) contextual additive structure and (B) ME-based context-dependent modeling.

In contrast, Figure 3B shows the corresponding ME-based modeling approach. In the previous subsection, it is described that ME-based context-dependent modeling needs two sets of regions, f l c l = 1 L f and g l c l = 1 L g . This example assumes that the leaves of Q1 and Q2 are defined as the first set of regions f l c l = 1 L f , and the leaves of Q3 are defined as the second set g l c l = 1 L g . Therefore, according to the explanation of the previous subsection, first empirical moments of Q1 and Q2, in addition to the second empirical moments of Q3, are captured by ME-based modeling. Figure 3B shows the estimated model mean values for all eight cubic clusters. As it is realized from the figure, model mean values estimated by ME-based modeling is a combination of adding parameters live in the regions f l c l = 1 L f divided by the parameters defined for the regions g l c l = 1 L g . In fact, the proposed ME-based modeling is an extension to the additive structure that ties all covariance matrices [19]. This extension is clear because if g l c l = 1 L g is defined with one region containing all contextual feature space, the ME-based modeling converts to the additive structure that ties all covariance matrices [19].

3.1.3 HMEM-based speech synthesis

HMEM improves both state duration distribution p i i = 1 N and output observation distribution b i i = 1 N using maximum entropy modeling. According to the discussion presented in Section 3.1.1, MEM requires two sets of contextual factors. In this section, for the sake of simplicity, it is assumed that the contextual regions defined for first-order moment constraints f l c l = 1 L f are identical to the regions defined for second-order moment constraints g l c l = 1 L g . All equations presented in this section is based on this assumption; however, their extension to the general case (different f l c l = 1 L f and g l c l = 1 L g ) is straightforward. Therefore, we define f l d i and f l o i as Ld and Lo contextual factors which are designed carefully for the purpose of modeling duration and acoustic parameters of the i th state. Maximum entropy criterion leads to the following duration and output probability distributions.

b i o t = g ϵ S o t w ig b i | g o t , b i | g o t = N n g o t ; 1 2 u ig H ig , H ig 1 , P i d = N 1 d ; 1 2 u i h i , 1 h i , u i = l = 1 L d f l d i u i l , h i = l = 1 L d f l d i h i l , u ig = l = 1 L o f l o i u ig l , H ig = l = 1 L o f l o i H ig l , w ig = exp l = 1 L o f l o i w ig l g = 1 G exp i = 1 L o f l o i w i g l .

In these equations, S(o t ) is a set of all possible spaces defined for o t . u i l and h i l are the duration model parameters, and w ig l , u ig l , and H ig l denote the output model parameters related to the l th contextual factor, g th space, and i th state.

We can now probe the differences between HSMM and HMEM context-dependent acoustic modeling. These two modeling approaches are dramatically close to each other, so that defining HMEM contextual factors based on the decision trees described by Equation 3 would reduce HMEM to HSMM. Accordingly, HMEM extends HSMM and enables its structure to exploit overlapped contextual factors.

Moreover, another significant conclusion that could be drawn from this section is that several HSMM concepts are transposable within the HMEM framework. These concepts involve Viterbi algorithm, methods which calculate forward/backward variables and occupation probabilities, and even all parameter generation algorithms [2628]. It just needs to define mean vectors, covariance matrices, and space probabilities of HSMM in accordance with Equation 20.

3.2 HMEM parameter re-estimation

In the training phase, we are given a set of K i.i.d. training data O k k = 1 K ; the goal is to find the best set of model parameters λ ^ , which maximizes the log likelihood:

λ ^ def ¯ ¯ argmax λ L λ , L λ def ¯ ¯ 1 K k = 1 K ln P O k | λ .

Substituting Equation 5 for the likelihood P(O(k)|λ) leads to an excessively complex optimization problem with seemingly impossible direct solution. The major issue is that the distribution wholly depends upon the latent variables which are unknown. The expectation maximization (EM) technique offers an iterative algorithm which overcomes this problem and accurately solves the issue:

λ n + 1 = argmax λ Q λ ; λ n , Q λ ; λ n = k all d , all q P d , q | O k ; λ n ln P O k , d , q | ; λ ,

where d and q represent possible state durations and space indexes for the k th training utterance and the second summation is calculated over all possible values of d and q. In general, these functions cannot be minimized in a closed-form expression. Therefore, a numerical optimization technique such as the Broyden-Fletcher-Goldfarb-Shanno (BFGS) [48] method or Newton algorithm has to be derived to find one of the local optima. This paper proposes to exploit the outstanding BFGS algorithm, due to its favorable characteristics. However, BFGS needs solely the first partial derivatives of the cost functions calculated as follows:

Q u i l = 1 2 k f l d i t = 1 T d = 1 t χ t d i d + u i 2 h i , Q h i l = 1 2 k f l d i t = 1 T d = 1 t χ t d i d 2 1 h i u i 2 h i 2 , Q w ig l = k f l o i t = 1 T γ t i , g 1 w ig , Q u ig l = 1 2 k f l o i t = 1 T γ t i , g o t + H ig 1 u ig 2 , Q H ig l = 1 2 k f l o i t = 1 T γ t i , g o t o t T H ig 1 H ig 1 u ig u ig T H ig 1 4 ,

where γ t (i, g) and χ t d i are defined in Section 2.3. Therefore, at every iteration of BFGS, we need to find the above gradient values and BFGS estimates new parameters which are closer to the optimum ones.

At first glance, calculating the above gradient expressions seems to be computationally expensive, but they can be calculated efficiently if we rewrite them in terms of sufficient statistics as in the following equations. By doing this, the computational complexity no longer depends on the number of training observation vectors, but rather on the total number of states. Furthermore, storing sufficient statistics instead of all observation vectors reduces the amount of main memory usage of the training procedure. These equations are expressed as

Q u i l = 1 2 k f l d i X ˜ i m ˜ i + u i 2 h i , Q h i l = 1 2 k f l d i X ˜ i r ˜ i 1 h i u i 2 h i 2 , Q w ig l = k f l o i γ ˜ i , g 1 w ig , Q u ig l = 1 2 k f l o i γ ˜ i , g μ ˜ i , g + H ig 1 u ig 2 , Q u ig l = 1 2 k f l o i γ ˜ i , g R ˜ i , g + H ig 1 u ig 2 ,

where X ˜ i , m ˜ i , and r ˜ i are sufficient statistics required to train duration distribution and are calculated as

X ˜ i = t = 1 T d = 1 t χ t d i , m ˜ i = 1 X ˜ i t = 1 T d = 1 t χ t d i d , r ˜ i = 1 X ˜ i t = 1 T d = 1 t χ t d i d 2 .

Also, γ ˜ i , g , μ ˜ i , g , and R ˜ i , g are sufficient statistics related to output probability distribution:

γ ˜ i , g = t = 1 T γ t i , g , μ ˜ 0 i , g = 1 γ ˜ i , g t = 1 T d = 1 t χ t d i o t , R ˜ i , g = 1 γ ˜ i , g t = 1 T d = 1 t χ t d i o t 2 .

These equations prove that regardless of calculating sufficient statistics, an EM iteration in HMEM is just equivalent to train three maximum entropy models for state duration distribution, state output distribution for each subspace, and subspace probability.Having introduced HMEM parameter estimation procedure, we can now proceed to explain the overall structure of HMEM. Figure 4 shows the whole architecture illustrating the HMEM-based speech synthesis system. Just like other statistical parametric approaches, it consists of two phases, training and synthesis. In the training phase, we first extract a parametric representation of the speech signal (i.e., acoustic features) including both spectral and excitation features from training speech database. In parallel, contextual factors are obtained for all states of the database. Thereafter, both acoustic and contextual factors are applied for HMEM training. The training procedure is performed by iterating through three steps: computing sufficient statistics, training all maximum entropy distributions, and calculating occupation probabilities. However, the training procedure needs prior information about state occupation probabilities for the first iteration. This paper proposes to utilize a trained HMM for this purpose. Training procedure continues until an amount of increase in likelihood falls below a specific threshold. The synthesis phase is completely identical to a typical HSMM-based speech synthesis system. The only difference is that in HMEM state, mean and covariance parameters are estimated in accordance with Equation 20 instead of tracing a binary decision tree.

Figure 4

Block-diagram of HMEM-based speech synthesis.

3.3 Decision tree-based context clustering

Statistical parametric speech synthesis systems typically exploit around 50 different types of contextual factors [23]. For such system, it is impossible to prepare tanning data covering all context-dependent models, and there are a large number of unseen models that have to be predicted in synthesis phase. Therefore, a context clustering approach such as decision tree-based clustering has to be used to decide about unseen contexts [31, 45]. Due to the critical importance of context clustering algorithms in HMM-based speech synthesis systems, this section focuses on designing a clustering algorithm for HMEM.

As it is realized from the discussion in this section, In order to implement the proposed architecture, we initially need to define two sets of contextual regions. These regions are represented by two sets, namely f l c l = 1 L f and g l c l = 1 L g . First- and second-order moment constraints have to be satisfied for all regions in f l c l = 1 L f and g l c l = 1 L g , respectively. Before training, the first empirical moments of all regions in f l c l = 1 L f and the second empirical moments of all regions in g l c l = 1 L g are computed using training data. Then, HMEM is trained to be consistent with these empirical moments. The major difficulty in defining these regions is to find a satisfactory balance between model complexity and the availability of training data. For limited training databases, a model with a small number of parameters, i.e., small number of regions has to be defined. In this case, bigger (strongly overlapped) contextual regions seem to be more desirable, because they can alleviate the problem of weak context generalization. On the other hand, for large training databases, larger number of contextual regions has to be defined to escape from under-fitting model to training data. In this case, smaller contextual regions can be applied to capture the details of acoustic features. This section introduces an algorithm that defines multiple contextual regions for first- and second-order moments by considering HMEM structure.

Due to the complex relationship between acoustic features and contextual factors, it is extremely difficult to find the optimum sets of contextual regions that maximize likelihood for HMEM. For the sake of simplicity, we have made some simplifying assumptions to find a number of suboptimum contextual regions. These assumptions are expressed as follows:

  • We have used conventional binary decision tree structures to define f l c l = 1 L f and g l c l = 1 L g . This is a common approach in many former papers [19, 20, 23]. It should be noted that the decision tree structure is not the only possible structure to express the relationship between acoustic features and contextual factors. For example, other approaches such as neural networks or soft-clustering methods can be applied as well. However, in this paper, we limit our discussion to the conventional binary decision tree structure.

  • Multiple decision trees are trained for f l c l = 1 L f , and just one decision tree is constructed for g l c l = 1 L g . In this way, the final HMEM preserves the first empirical moments of multiple decision trees, and the second moments of just one decision tree. This assumption is a result of the fact that first-order moments seem to be more important than second-order moments [32, 47].

  • The discussion of current section shows that the ML estimates of parameters defined for f l c l = 1 L f and g l c l = 1 L g significantly depend on each other. Therefore, in each step of decision tree construction, a BFGS optimization algorithm has to be executed to re-estimate both sets of parameters simultaneously, and this procedure leads to an extreme amount of computational complexity. To alleviate this problem, it is proposed to borrow g l c l = 1 L g from a baseline system (conventional HMM-based speech synthesis system) and construct f l c l = 1 L f independently.

  • In HMEM structure, f l c l = 1 L f is responsible to provide satisfactory clustering of first-order moments (mean vectors). Similarly, contextual additive structures [19, 20, 37] that tie all covariance matrices offer multiple overlapped clustering of mean vectors based on the likelihood criterion; therefore, an appropriate method is to borrow f l c l = 1 L f from the contextual additive structure.

  • However, training a contextual additive structure using algorithms proposed in [19, 20] is still computationally expensive for large training databases (more than 500 sentences). Three modifications are applied to the algorithm proposed by Takaki et al. [19] for computational complexity reduction: (i) The number of decision trees is considered to be fixed (in our experiments, an additive structure with four decision trees is built). (ii) Questions are selected one by one for different decision trees. Therefore, all trees are grown simultaneously, and the size of all trees would be equal. (iii) In the process of selecting the best pair of question and leaf, it is assumed that just the parameters of candidate leaf will be changed and all other parameters remain unchanged. It should be noted that the selection procedure is repeated until the total number of free parameters reaches the number of parameters trained for the baseline system (HSMM-based speech synthesis system).

In sum, the final algorithm of determining f l c l = 1 L f and g l c l = 1 L g can be summarized as follows. g l c l = 1 L g is simply borrowed from a conventional HMM-based speech synthesis system. f l c l = 1 L f also resulted from an independent context clustering algorithm that is a fast and simplified version of contextual additive structure [19]. This clustering algorithm builds four binary context-dependent decision trees, simultaneously. It should be noted that when the number of clusters reaches the number of leaves of the decision tree trained for an HSMM-based system, the clustering algorithm is finished.

The following algorithm shows the overall procedure of the proposed context clustering.

4 Experiments

We have conducted two sets of experiments. First, the performance of HMEM with heuristic context clusters is examined; second, the impact of the proposed method for decision tree-based context clustering presented in the Section 3.3 is evaluated.

4.1 Performance evaluation of HMEM with heuristic context clusters

This subsection aims to compare HMEM-based acoustic modeling with conventional HSMM-based method. In this subsection, contextual regions of HMEM are defined heuristically and it is fixed for different sizes of training database.

4.1.1 Experimental conditions

A Persian speech database [49] consisting of 1,000 utterances from a male speaker was used throughout our experiments. Sentences were between 5 and 20 words long and have an average duration of 8 s. This database was specifically designed for the purpose of speech synthesis. Sentences in the database covered most frequent Persian words, all bi-letter combinations, all bi-phoneme combinations, and most frequent Persian syllables. In the modeling of the synthesis units, 31 phonemes were used, including silence. As presented in Section 4.1.2, a large variety of phonetic and linguistic contextual factors was considered in this work.

Speech signals were sampled at a rate of 16 kHz and windowed by a 25-ms Blackman window with a 5-ms shift. 40 Mel-cepstral coefficients, 5 bandpass aperiodicity and fundamental frequency, and their delta and delta-delta coefficients extracted by STRAIGHT [11] were employed as our acoustic features. In this experiment, the number of states was 5, and multi-stream left-to-right with no skip path MSD-HSMM was trained as the traditional HSMM system. Decision trees were built using maximum likelihood criterion, and the size of decision trees was determined by MDL principle [46]. Additionally, global variance (GV)-based parameter generation algorithm [20, 26] and STRAIGHT vocoder were applied in the synthesis phase.

Both subjective and objective tests were carried out to compare HMEM that uses some heuristic contextual regions with the traditional HSMM system. In our experiments, two different synthesis systems named HMEM1 and HMEM2 were developed based on the proposed approach. HMEM1 employs a small number of general highly overlapped contextual factors that are designed carefully for each stream, while HMEM2 uses a larger number of contextual factors.

More precisely, a set of 64 initial contextual factors were extracted for each segment (phoneme) of the Persian database. These factors contain both segmental and suprasegmental contextual features. From these contextual factors, a set of approximately 8,000 contextual questions were designed and the HSMM system was trained using these questions. Each question can form two regions; therefore, these 8,000 questions can be converted to 16,000 regions. For each stream of HMEM1, a small number of these contextual regions that seem to be more important for that stream were selected and HMEM1 was trained using them. Contextual factors of HMEM2 contain all contextual factors of HMEM1 in addition to a number of detailed ones. The number of contextual regions in HMEM2 is twice the number of regions in HMEM1. Regions of both HMEM1 and HMEM2 were selected based on the linguistic knowledge of the Persian language. Table 1 shows the number of contextual regions for different synthesis systems (namely HSMM with different training data sizes, HMEM1 and HMEM2).

Table 1 The number of leaf nodes for each stream in different speech synthesis systems

Experiments were conducted on five different training sets with 50, 100, 200, 400, and 800 utterances. Additionally, a fixed set of 200 utterances, not included in the training sets, was used for testing.

4.1.2 Employed contextual factors

In our experiments, contextual factors contained phonetic, syllable, word, phrase, and sentence level features. In each of these levels, both general and detailed features were considered. Features such as phoneme identity, syllable stress pattern, or word part-of-speech tag are examples of general features, and a question like the position of the current phoneme is a sample of a detailed one. Specific information with regard to contextual features is presented in this subsection.

Contextual factors play a significant role in the proposed HMEM method. As a consequence, they have been designed carefully and are now briefly presented:

  • ➢ Phonetic-level features

    • Phoneme identity before the preceding phoneme; preceding, current, and succeeding phonemes; and phoneme identity after the next phoneme

    • Position of the current phoneme in the current syllable (forward and backward)

    • Whether this phoneme is ‘Ezafe’ [50] or not (Ezafe is a special feature in Persian pronounced as a short vowel ‘e’ and relates two different words together. Ezafe is not written but is pronounced and has a profound effect on intonation)

  • ➢ Syllable-level features

    • Stress level of this syllable (five different stress levels are defined for our speech database)

    • Position of the current syllable in the current word and phrase (forward and backward)

    • Type of the current syllable (syllables in Persian language are structured as CV, CVC, or CVCC, where C and V denote consonants and vowels, respectively)

    • Number of the stressed syllables before and after the current syllable in the current phrase

    • Number of syllables from the previous stressed syllable to the current syllable

    • Vowel identity of the current syllable

  • ➢ Word-level features

    • Part-of-speech (POS) tag of the preceding, current and succeeding word

    • Position of the current word in the current sentence (forward and backward)

    • Whether the current word contains ‘Ezafe’ or not

    • Whether this word is the last word in the sentence or not

  • ➢ Phrase-level features

    • Number of syllables in the preceding, current, and succeeding phrase

    • Position of the current phrase in the sentence (forward and backward)

  • ➢ Sentence-level features

    • Number of syllables, words, and phrases in the current sentence

    • Type of the current sentence

4.1.3 Illustratory example

Before going further with the objective and subjective evaluations, the superiority of HMEM over HSMM when few training data are available can be already illustrated. Although the improvement will be shown in Sections 4.1.4 and 4.1.5 to be achieved for all speech characteristics (log F0, duration, and spectral features), it is here emphasized for the prediction of log F0 trajectories. Figure 5 shows the trajectory of log F0 generated by HSMM and HMEM1 with 100 training utterances, in contrast to the natural contour. This plot confirms the superiority of HMEM over HSMM in modeling fundamental frequency when the amount of training data is small, as the generated contour by HMEM is far closer to the natural one compared to HSMM.

Figure 5

Trajectory of log F0 generated from the HSMM, HMEM as well as the natural log F0.

In limited training sets, HSMM produces sudden transitions between adjacent states. This drawback is the result of decision tree-clustered context-dependent modeling. More specifically, when few data are available for training, the number of leaves in the decision tree is reduced. As a result, the distance between the mean vectors of adjacent states can be large. Even the parameter generation algorithm proposed by [2628] cannot compensate such jumps. In such cases, the quality of synthetic speech with HSMM is expected to deteriorate.

On the opposite, if we let adjacent states contain common active contextual factors, then the variation of mean vectors in state transitions will be smoother. This is the key idea of HMEM which makes it possible to outperform HSMM when the data are limited. However, the use of overlapped contextual factors in HMEM will result in over-smoothing problem when the size of the training data is increased. Therefore, the detailed contextual factors are additionally considered in HMEM2 to alleviate the over-smoothing issue.

4.1.4 Objective evaluation

The average mel-cepstral distortion (MCD) [51] and root-mean-square (RMS) error of phoneme durations (expressed in terms of number of frames) were selected as relevant metrics for our objective assessment. For the calculation of both average mel-cepstral distance and RMS error of phoneme durations, the state boundaries (state durations) were determined using Viterbi alignment with the speaker's real utterance.

The MCD measure is defined by:

MCD = 10 ln 10 * 2 i = 1 40 m c i t m c i p 2 ,

where mc i is the i th mel-cepstral coefficients in a frame, mct is the target coefficient we are comparing against, and mcp is the generated coefficient. In addition, RMS is defined as the following function:

RMS = s = 1 N d s t d s p 2 / N ,

where N is the total number of states in a sentence, d s is the duration of the s th state, d s t is the original duration, and d s p is the estimated duration.Figure 6 shows the average mel-cepstral distance between spectra generated from the proposed method and spectra obtained by analyzing the speaker's real utterance. For comparison, we also present the average distance of spectra generated from the HSMM-based method and the real spectra. In this figure, it is clearly observed that the proposed HMEM systems outperform the standard HSMM approach for limited training datasets. Nonetheless, this advantage disappears when more than 200 utterances are available for training. It can be noticed that a reduction of the size of the training set has a dramatic impact on the performance of HSMM, contrary to HMEM-based systems.The same conclusions are observed for Figure 7 in which the generated duration of proposed systems is compared against that of HSMM. It can be again noticed that the proposed systems outperform HSMM in small databases. However, when the size of the database increases, HSMM gradually surpasses the proposed HMEM systems. Furthermore, detailed features added in HMEM2 affect the proposed method constructively when the synthesis units model by large databases. Thus, we expect that the proposed method could be comparable with HSMM or outperform it even for large databases if we apply more detailed and well-designed features.

Figure 6

Comparison of average MCD as an objective measure between the proposed method and the HSMM-based one.

Figure 7

Comparison of RMS error of phoneme durations as objective measure between the proposed method and HSMM one.

In summary, from these figures and the illustratory example presented before, we can see that when the available data are limited, all features (log F0, duration, and spectra) of synthetic speech generated by HMEM are closer to the original features than those obtained with HSMM. However, when the training database is large, the HSMM-based method performs better than HMEM. Nevertheless, employing more detailed features can assist the proposed method in becoming closer to the HSMM-based synthetic speech.

In addition to the abovementioned objective measurements, we have compared the accuracy of voiced/unvoiced detection in the proposed system with its counterpart in HSMM-based synthesis. Table 2 shows information about the false negative (FN), false positive (FP), true negative (TN), and true positive (TP) rates. Moreover, the data in Table 2 are summarized in Table 3 in which the accuracy of detecting voice/unvoiced regions is presented. As realized from these tables, the proposed method detects voiced/unvoiced regions more accurately than HSMM regardless of the size of the database. In other words, not only in small databases but also for larger ones, HMEM outperforms HSMM in terms of detecting voiced/unvoiced regions.

Table 2 FN, FP, TN, and TP rates of detecting voiced/unvoiced regions through HMEM2 and the HSMM-based method
Table 3 Accuracy of voiced/unvoiced detector

4.1.5 Subjective evaluation

Two different subjective methods are employed in order to show the effectiveness of the proposed system and assess the effect of the size of the training database. A comparative mean opinion score (CMOS) test [52] with a 7-point scale, ranging from −3 (meaning that method A is much better than method B) to 3 (meaning the opposite), and a preference scoring [53] are used to evaluate the subjective quality of the synthesized speech. The results of this evaluation are respectively shown in Figures 8 and 9.

Figure 8

Averaged CMOS scores for the HMEM1, HMEM2, and HSMM. 95% confidence intervals are also indicated.

Figure 9

Preference scores as a function of the number of utterances used for training. (A) Comparison between HMEM1 and HMEM2. (B) Comparison between HMEM1 and HSMM. (C) Comparison between HMEM2 and HSMM.

Twenty native participants were asked to listen to ten randomly chosen pairs of synthesized speech samples generated by two different systems (selected arbitrarily among HMEM1, HMEM2, and HSMM).

Remarkably, the proposed systems are noticed to be of a great interest when the training data are limited (i.e., for 50, 100, and 200 utterances) and are in line with the conclusions of the objective assessments. The superiority of HMEM1 over HSMM and HMEM2 is clear in the training sets containing 50 and 100 utterances. In other words, general contextual factors lead the proposed system to a better performance when the amount of training data is very small. Gradually, as the number of utterances in the training set increases, detailed features assist the proposed system in achieving more effective synthetic speech. Therefore, HMEM2 surpasses HMEM1 for training sets with 200 and more utterances. However, for relatively large training sets (400 and 800), the use of HSMM is recommended.

Table 1 compares the number of leaf nodes in different speech synthesis systems. It can be seen from the table that to model mgc stream, HMEM2 exploits more parameters than HSMM-400 and HSMM-800, but the objective evaluations presented in Figure 6 show that HSMM-400 and HSMM-800 results in better mel-cepstral distances. The above argument shows that HMEM with some heuristic contextual clusters cannot exploit model parameters efficiently. In fact, a great number of contextual regions in HMEM1 and HMEM2 are redundant; therefore, their corresponding parameters are not useful. The next section evaluates the performance of HMEM with the suboptimum context clustering algorithm proposed in Section 3.3. This proposed clustering algorithm selects appropriate contextual regions and consequently solves the aforementioned problem.

4.2 Performance evaluation of HMEM with decision tree-based context clustering

This section is dedicated to the second set of experiments conducted to evaluate the performance of HMEM with decision tree construction algorithm proposed in Section 3.3. As it is realized from the first set of experiments, HMEM with heuristic and naïve contextual regions cannot outperform HSMM in large training databases. This section proves that by employing appropriate sets of f l c l = 1 L f and g l c l = 1 L g , HMEM outperforms HSMM even for large databases.

4.2.1 Experimental conditions

Experiments were carried out on Nick [54], a British male database collected in Edinburgh University. This database consists of 2,500 utterances from a male speaker. We considered five sets including 50, 100, 200, 400, and 800 utterances for training, and 200 sentences that were not included in training sets were used as test data. Each sentence in the database is about 5 s of speech. Speech signals are sampled at 48 kHz, windowed by a 25-ms Blackman window with 5-ms shift. This database was specifically designed for the purpose of speech synthesis research, and utterances in the database covered most frequent English words. Also, different segmental and suprasegmental contextual factors were extracted for this database.

The speech analysis conditions and model topologies of CSTR/EMIME HTS 2010 [54] were used in this experiment. Bark cepstrum was extracted from smooth STRAIGHT trajectories [11]. Also, instead of log F0 and five frequency sub-bands (0 to 1, 1 to 2, 2 to 4, 4 to 6, and 6 to 8 kHz), pitch in mel and auditory-scale motivated frequency bands for aperiodicity measure were applied [54]. The analysis process resulted in 40 bark cepstrum coefficients, 1 mel in pitch value, and 25 auditory-scale motivated frequency bands aperiodicity parameters for each frame of training speech signals. These parameters incorporated with their delta and delta-delta parameters considered as the observation vectors of the statistical parametric model.

A five-state multi-stream left-to-right with no skip path MSD-HSMM was trained as the baseline system. Conventional maximum likelihood-based decision tree clustering algorithm was used to tie HMM states, but MDL criterion is used to determine the size of decision trees.

In order to have a fair comparison, the proposed system (HMEM with decision tree structure) was trained with the same number of free model parameters as the baseline system. HMEM was trained based on the decision tree construction algorithm presented in Section 3.3 and parameter re-estimation algorithm proposed in Section 3.2. It should be noted that four decision trees were built for f l c l = 1 L f and one decision tree for g l c l = 1 L g . After training acoustic models, in the synthesis phase, GV-based parameter generation algorithm [20, 26] and STRAIGHT synthesis module generated synthesized speech signals. Both subjective and objective tests were conducted to compare HMEM that uses decision tree-based clusters with traditional HSMM-based synthesis.

It is useful to mention that training the proposed HMEM structure with decision tree-based context clustering took approximately 5 days for 800 training sentences, while training its corresponding HSMM-based synthesis system took approximately 16 h.

4.2.2 Employed contextual factors

In this experiment, employed contextual factors contained phonetic, syllable, word, phrase, and sentence level factors. In each of these levels, all important features were considered. Specific information about these features is presented in this subsection.

  • ➢ Phonetic-level features

    • Phoneme identity before the preceding phoneme; preceding, current, and succeeding phonemes; and phoneme identity after the next phoneme

    • Position of the current phoneme in the current syllable, word, phrase, and sentence

  • ➢ Syllable-level features

    • Stress level of previous, current, and next syllable (three different stress levels are defined for this database)

    • Position of the current syllable in the current word, phrase, and sentence

    • Number of the phonemes of the previous, current, and next syllable

    • Whether the previous, current, and next syllable is accented or not

    • Number of the stressed syllables before and after the current syllable in the current phrase

    • Number of syllables from the previous stressed syllable to the current syllable

    • Number of syllables from the previous accented syllable to the current syllable

  • ➢ Word-level features

    • Part-of-speech (POS) tag of the preceding, current, and succeeding word

    • Position of the current word in the current phrase and sentence (forward and backward)

    • Number of syllables of the previous, current, and next word

    • Number of content words before and after current word in the current phrase

    • Number of words from previous and next content word

  • ➢ Phrase-level features

    • Number of syllables and words of the preceding, current, and succeeding phrase

    • Position of the current phrase in the sentence

    • Current phrase ToBI end tone

  • ➢ Sentence-level features

    • Number of phonemes, syllables, words, and phrases in the current utterance

    • Type of the current sentence

4.2.3 Objective evaluation

Two well-known measures were applied for objective evaluation of the proposed decision tree-based HMEM in comparison with conventional HSMM. The first measure computes RMS error of generated log F0 trajectories, and the second one compares synthesized spectrograms using average MCD criterion. The results of these measures are shown in Figures 10 and 11. As it is realized from Figure 10 that shows the RMS error of the log F0 in terms of cent for different sizes of training data, the log F0 trajectories generated from the proposed approach are more similar to the natural log F0 trajectories, and therefore, HMEM improves the performance of log F0 modeling. However, by increasing the size of the database, the amount of this improvement is slightly reduced. Hence, it can be implied from this figure that in log F0 modeling, the effect of applying overlapped regions for small databases is relatively more than its effect on big databases. Additionally, Figure 11 shows the result of average MCD test. This result also confirms the improvement of HMEM performance in contrast to conventional HSMM for all training databases. As it is clear from the figure, the improvement in average MCD test is fixed for all databases.

Figure 10

RMSE as objective measure to compare log F0 trajectories generated by decision tree-based HMEM and conventional HSMM.

Figure 11

Result of the MCD measure that compares decision tree-based HMEM and conventional HSMM.

4.2.4 Subjective evaluation

We conducted paired comparison tests and reported comparative mean opinion score (CMOS) and preference score as subjective evaluations. Fifteen non-professional native listeners were presented with 30 randomly chosen pairs of synthesized speech generated by HMEM and HSMM. Listeners selected the synthesized speech which sounds better and determined how much is better (much better, better, slightly better, or about the same). The results are shown in Figures 12 and 13.

Figure 12

Subjective evaluation of HMEM with decision tree-based context clustering and HSMM through CMOS test with 95% confidence intervals.

Figure 13

Preference scores as a subjective comparison between HMEM with decision tree-based context clustering and HSMM.

Both CMOS test and preference score confirm the superiority of the proposed method over HSMM in all databases. Thus, if context clusters are determined through an effective approach, the proposed HMEM will outperform HSMM.

5. Conclusions

This paper addressed the main shortcomings of HSMM in context-dependent acoustic modeling, namely inadequate context generalization. HSMM uses decision tree-based context clustering that does not provide efficient generalization, because each acoustic feature vector is associated with modeling only one context cluster. In order to alleviate this problem, this paper proposed HMEM as a new acoustic modeling technique based on maximum entropy modeling approach. HMEM improves HSMM by enabling its structure to take advantage of overlapped contextual factors, and therefore, it can provide superior context generalization. Experimental results using objective and subjective criteria showed that the proposed system outperforms HSMM.

Despite the advantages, which enabled our system to outperform HSMM, a drawback of computationally complex training procedure is noticed in large databases.


  1. 1.

    Zen H, Tokuda K, Black AW: Statistical parametric speech synthesis. Speech Comm. 2009, 51(11):1039-1064. 10.1016/j.specom.2009.04.004

    Article  Google Scholar 

  2. 2.

    Black AW, Zen H, Tokuda K: Statistical parametric speech synthesis, in IEEE International Conference on Acoustics, vol. 4. Speech and Signal Processing (ICASSP), Honolulu, Hawaii, USA; 2007:IV1229-IV1232.

    Google Scholar 

  3. 3.

    Yamagishi J, Kobayashi T: Average-voice-based speech synthesis using HSMM-based speaker adaptation and adaptive training. IEICE - Trans. Info. Syst. 2007, 90(2):533-543.

    Article  Google Scholar 

  4. 4.

    Yamagishi J, Nose T, Zen H, Ling ZH, Toda T, Tokuda K, King S, Renals S: Robust speaker-adaptive HMM-based text-to-speech synthesis, In IEEE Transactions on Audio, Speech, and Language Processing. 2009, 17(6):1208-1230.

    Google Scholar 

  5. 5.

    Yamagishi J, Kobayashi T, Nakano Y, Ogata K, Isogai J: Analysis of speaker adaptation algorithms for HMM-based speech synthesis and a constrained SMAPLR adaptation algorithm, IEEE Transactions on Audio, Speech, and Language Processing. 2009, 17(1):66-83.

    Google Scholar 

  6. 6.

    Wu YJ, Nankaku Y, Tokuda K: State mapping based method for cross-lingual speaker adaptation in HMM-based speech synthesis. In INTERSPEECH. Brighton, UK; 2009:528-531.

    Google Scholar 

  7. 7.

    Liang H, Dines J, Saheer L: A comparison of supervised and unsupervised cross-lingual speaker adaptation approaches for HMM-based speech synthesis. In IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP). Dallas, Texas, USA; 2010:4598-4601.

    Google Scholar 

  8. 8.

    Gibson M, Hirsimaki T, Karhila R, Kurimo M, Byrne W: Unsupervised cross-lingual speaker adaptation for HMM-based speech synthesis using two-pass decision tree construction. In IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP). Dallas, Texas, USA; 2010:4642-4645.

    Google Scholar 

  9. 9.

    Yamagishi J, Ling Z, King S: Robustness of HMM-based speech synthesis. In INTERSPEECH. Brisbane, Australia; 2008:581-584.

    Google Scholar 

  10. 10.

    Yoshimura T, Tokuda K, Masuko T, Kobayashi T, Kitamura T: Mixed excitation for HMM-based speech synthesis. In INTERSPEECH. Aalborg, Denmark; 2001:2263-2266.

    Google Scholar 

  11. 11.

    Kawahara H, Masuda-Katsuse I, de Cheveigné A: Restructuring speech representations using a pitch-adaptive time–frequency smoothing and an instantaneous-frequency-based F0 extraction: possible role of a repetitive structure in sounds. Speech Comm. 1999, 27(3):187-207.

    Article  Google Scholar 

  12. 12.

    Drugman T, Wilfart G, Dutoit T: A deterministic plus stochastic model of the residual signal for improved parametric speech synthesis. In INTERSPEECH. Brighton, United Kingdom; 2009:1779-1782.

    Google Scholar 

  13. 13.

    Drugman T, Dutoit T: The deterministic plus stochastic model of the residual signal and its applications. IEEE Trans. Audio. Speech. Lang. Process 2012, 20(3):968-981.

    Article  Google Scholar 

  14. 14.

    Zen H, Tokuda K, Masuko T, Kobayasih T, Kitamura T: A hidden semi-Markov model-based speech synthesis system. IEICE - Trans. Info. Syst 2007, 90(5):825.

    Article  Google Scholar 

  15. 15.

    Hashimoto K, Nankaku Y, Tokuda K: A Bayesian approach to hidden semi Markov model based speech synthesis, in Proceedings of INTERSPEECH. Brighton, United Kingdom; 2009:1751-1754.

    Google Scholar 

  16. 16.

    Hashimoto K, Zen H, Nankaku Y, Masuko T, Tokuda K: A Bayesian approach to HMM-based speech synthesis, in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Taipei, Taiwan; 2009:4029-4032.

    Google Scholar 

  17. 17.

    Tokuda K, Masuko T, Miyazaki N, Kobayashi T: Multi-space probability distribution HMM. IEICE Trans. on Info. Syst 2002, 85(3):455-464.

    Google Scholar 

  18. 18.

    Zen H, Senior A, Schuster M: Statistical parametric speech synthesis using deep neural networks. In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Vancouver, British Columbia, Canada; 2013:7962-7966.

    Google Scholar 

  19. 19.

    Takaki S, Nankaku Y, Tokuda K: Spectral modeling with contextual additive structure for HMM-based speech synthesis. In Proceedings of 7th ISCA Speech Synthesis Workshop. Kyoto, Japan; 2010:100-105.

    Google Scholar 

  20. 20.

    Takaki S, Nankaku Y, Tokuda K: Contextual partial additive structure for HMM-based speech synthesis. In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Vancouver, British Columbia, Canada; 2013:7878-7882.

    Google Scholar 

  21. 21.

    Gales MJ: Cluster adaptive training of hidden Markov models. IEEE Trans. Speech. Audio. Process. 2000, 8(4):417-428. 10.1109/89.848223

    Article  Google Scholar 

  22. 22.

    Zen H, Gales MJ, Nankaku Y, Tokuda K: Product of experts for statistical parametric speech synthesis, IEEE Trans. Audio. Speech. Lang. Process. 2012, 20(3):794-805.

    Article  Google Scholar 

  23. 23.

    Yu K, Zen H, Mairesse F, Young S: Context adaptive training with factorized decision trees for HMM-based statistical parametric speech synthesis. Speech Comm. 2011, 53(6):914-923. 10.1016/j.specom.2011.03.003

    Article  Google Scholar 

  24. 24.

    Toda T, Young S: Trajectory training considering global variance for HMM-based speech synthesis. In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Taipei, Taiwan; 2009:4025-4028.

    Google Scholar 

  25. 25.

    Qin L, Wu YJ, Ling ZH, Wang RH, Dai LR: Minimum generation error criterion considering global/local variance for HMM-based speech synthesis. In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Las Vegas, Nevada, USA; 2008:4621-4624.

    Google Scholar 

  26. 26.

    Toda T, Tokuda K: Speech parameter generation algorithm considering global variance for HMM-based speech synthesis. IEICE - Trans. Info. Syst. Arch 2007, E90-D(5):816-824. 10.1093/ietisy/e90-d.5.816

    Article  Google Scholar 

  27. 27.

    Tokuda K, Yoshimura T, Masuko T, Kobayashi T, Kitamura T: Speech Parameter Generation Algorithms for HMM-based Speech Synthesis, in ICASSP, vol. 3. Istanbul; 2000:1315-1318.

    Google Scholar 

  28. 28.

    Tokuda K, Kobayashi T, Imai S: Speech parameter generation from HMM using dynamic features. In International Conference on Acoustics, Speech, and Signal Processing (ICASSP), vol. 1. Detroit, Michigan, USA; 1995:660-663.

    Google Scholar 

  29. 29.

    Comparing glottal-flow-excited statistical parametric speech synthesis methods In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Vancouver, British Columbia, Canada; 2013:7830-7834.

  30. 30.

    Yoshimura T, Tokuda K, Masuko T, Kobayashi T, Kitamura T: Simultaneous modeling of spectrum, pitch and duration in HMM-based speech synthesis. Proceedings of Eurospeech 1999, 2347-2350.

    Google Scholar 

  31. 31.

    Young SJ, Odell JJ, Woodland PC: Tree-based state tying for high accuracy acoustic modeling. Proceedings of the workshop on Human Language Technology, Association for Computational Linguistics 1994, 307-312.

    Google Scholar 

  32. 32.

    Leggetter CJ, Woodland PC: Maximum likelihood linear regression for speaker adaptation of continuous density hidden Markov models. Comput. Speech. Lang. 1995., 9(2):

    Google Scholar 

  33. 33.

    Digalakis VV, Neumeyer LG: Speaker adaptation using combined transformation and Bayesian methods. IEEE Trans. Speech. Audio. Process. 1996, 4(4):294-300. 10.1109/89.506933

    Article  Google Scholar 

  34. 34.

    Zen H, Braunschweiler N, Buchholz S, Gales MJ, Knill K, Krstulovic S, Latorre J: Statistical parametric speech synthesis based on speaker and language factorization. IEEE Transactions. Audio. Speech. Lang. Process. 2012, 20(6):1713-1724.

    Article  Google Scholar 

  35. 35.

    Koriyama T, Nose T, Kobayashi T: Statistical parametric speech synthesis based on Gaussian process regression, IEEE Journal of Selected Topics in Signal Processing. 2013, 1-11.

    Google Scholar 

  36. 36.

    Yu K, Mairesse F, Young S: Word-level emphasis modeling in HMM-based speech synthesis. In IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP). Dallas, Texas, USA; 2010:4238-4241.

    Google Scholar 

  37. 37.

    Zen H, Braunschweiler N: Context-dependent additive log f_0 model for HMM-based speech synthesis. In INTERSPEECH. Brighton, United Kingdom; 2009:2091-2094.

    Google Scholar 

  38. 38.

    Sakai S: Additive modeling of English f0 contour for speech synthesis, in Proceedings of ICASSP. Las Vegas, Nevada, USA; 2008:277-280.

    Google Scholar 

  39. 39.

    Qian Y, Liang H, Soong FK: Generating natural F0 trajectory with additive trees. In INTERSPEECH. Brisbane, Australia; 2008:2126-2129.

    Google Scholar 

  40. 40.

    Wu YJ, Soong F: Modeling pitch trajectory by hierarchical HMM with minimum generation error training. In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Kyoto, Japan; 2012:4017-4020.

    Google Scholar 

  41. 41.

    Berger AL, Pietra VJD, Pietra SAD: A maximum entropy approach to natural language processing. Computer Ling 1996, 22: 39-71. 10.1016/0096-0551(96)00005-7

    Article  Google Scholar 

  42. 42.

    Borthwick A: A maximum entropy approach to named entity recognition, PhD dissertation (New York University). 1999.

    Google Scholar 

  43. 43.

    Rangarajan V, Narayanan S, Bangalore S: Exploiting acoustic and syntactic features for prosody labeling in a maximum entropy framework, in Proceedings of NAACL HLT. 2007, 1-8.

    Google Scholar 

  44. 44.

    Ratnaparkhi A: A maximum entropy model for part-of-speech tagging, in Proceedings of the conference on empirical methods in natural language processing. 1996, 1: 133-142.

    Google Scholar 

  45. 45.

    Odell JJ: The use of context in large vocabulary speech recognition, PhD dissertation (Cambridge University). 1995.

    Google Scholar 

  46. 46.

    Shinoda K, Takao W: MDL-based context-dependent subword modeling for speech recognition. J. Acoust. Soc. Jpn 2000, 21(2):79-86. 10.1250/ast.21.79

    Article  Google Scholar 

  47. 47.

    Oura K, Zen H, Nankaku Y, Lee A, Tokuda K: A covariance-tying technique for HMM-based speech synthesis. J. IEICE 2010, E93-D(3):595-601.

    Google Scholar 

  48. 48.

    Nocedal J, Stephen JW: Numerical Optimization. Book of Springer, USA; 1999.

    Book  Google Scholar 

  49. 49.

    Bijankhan M, Sheikhzadegan J, Roohani MR, Samareh Y, Lucas C, Tebiani M: The speech database of Farsi spoken language. Proceedings of 5th Australian International Conference on Speech Science and Technology (SST) 1994, 826-831.

    Google Scholar 

  50. 50.

    Ghomeshi J: Non-projecting nouns and the ezafe: construction in Persian. Nat. Lang. Ling. Theor. 1997, 15(4):729-788. 10.1023/A:1005886709040

    Article  Google Scholar 

  51. 51.

    Kubichek R: Mel-cepstral distance measure for objective speech quality assessment. IEEE Pacific Rim Conference on Communications, Computers and Signal Processing, vol. 1 1993, 125-128.

    Chapter  Google Scholar 

  52. 52.

    Picart B, Drugman T, Dutoit T: Continuous control of the degree of articulation in HMM-based speech synthesis, 12th Annual Conference of the International Speech Communication Association (ISCA). INTERSPEECH, Florence, Italy; 2011:1797-1800.

    Google Scholar 

  53. 53.

    Yamagishi J: Average-Voice-Based Speech Synthesis, PhD dissertation. Tokyo Institute of 1362 Technology, Yokohama; 2006.

    Google Scholar 

  54. 54.

    Yamagishi J, Watts O: The CSTR/EMIME HTS system for Blizzard challenge, in Proceedings of Blizzard Challenge 2010. Kyoto, Japan; 2010:1-6.

    Google Scholar 

Download references

Author information



Corresponding author

Correspondence to Soheil Khorram.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ original submitted files for images

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and Permissions

About this article

Cite this article

Khorram, S., Sameti, H., Bahmaninezhad, F. et al. Context-dependent acoustic modeling based on hidden maximum entropy model for statistical parametric speech synthesis. J AUDIO SPEECH MUSIC PROC. 2014, 12 (2014).

Download citation


  • Hidden Markov model (HMM)-based speech synthesis
  • Context-dependent acoustic modeling
  • Decision tree-based context clustering
  • Maximum entropy
  • Overlapped context clusters
  • Statistical parametric speech synthesis