- Research
- Open Access

# Regularized minimum class variance extreme learning machine for language recognition

- Jiaming Xu
^{1, 2}Email author, - Wei-Qiang Zhang
^{3}, - Jia Liu
^{3}and - Shanhong Xia
^{1}

**2015**:22

https://doi.org/10.1186/s13636-015-0066-5

© Xu et al. 2015

**Received:**25 November 2014**Accepted:**3 August 2015**Published:**13 August 2015

## Abstract

Support vector machines (SVMs) have played an important role in the state-of-the-art language recognition systems. The recently developed extreme learning machine (ELM) tends to have better scalability and achieve similar or much better generalization performance at much faster learning speed than traditional SVM. Inspired by the excellent feature of ELM, in this paper, we propose a novel method called regularized minimum class variance extreme learning machine (RMCVELM) for language recognition. The RMCVELM aims at minimizing empirical risk, structural risk, and the intra-class variance of the training data in the decision space simultaneously. The proposed method, which is computationally inexpensive compared to SVM, suggests a new classifier for language recognition and is evaluated on the 2009 National Institute of Standards and Technology (NIST) language recognition evaluation (LRE). Experimental results show that the proposed RMCVELM obtains much better performance than SVM. In addition, the RMCVELM can also be applied to the popular i-vector space and get comparable results to the existing scoring methods.

## Keywords

- Language recognition
- Extreme learning machine
- Single-hidden layer feedforward neural networks
- Support vector machine

## 1 Introduction

The task of spoken language recognition is to automatically determine the identity of the language spoken in a speech sample [1]. It has received increasing attention due to its importance for the enhancement of automatic speech recognition (ASR) [2] and multi-language translation systems [3, 4].

The most popular language recognition techniques can usually be categorized as phonotactic [5–7] and acoustic approaches [8, 9]. The former method is based upon phone recognizer followed by language models (PRLM), parallel PRLM (PPRLM), and supper vector machines PRLM. The later approach uses shifted-delta-cepstral (SDC) coefficients as a means of incorporating temporal information about the speech into the feature vector.

The introduction of support vector machines (SVMs) into language recognition resulted in significant improvement in performance [10]. SVMs have proven to be an effective method and have been widely used for many years [11]. SVMs perform a nonlinear mapping from an input space to an SVM feature space. The standard optimization method is then used to find the solution of maximizing the separating margin of two different classes in this potentially high-dimensional feature space while minimizing the training errors. In language recognition, we usually adopt the one-versus-rest tactic to deal with multiclass classification problems when using SVMs, namely, setting data from target language as the positive samples and data from all the other languages as the negative samples. In this way, the amount of training samples between the positive language and negative languages is unbalanced and sometimes it is even impossible to separate them properly by a linear hyperplane. In addition, the training of SVMs involves a quadratic programming (QP) problem, the computational complexity of SVM training algorithms is usually intensive, which is at least quadratic with respect to the number of training samples.

In recent years, extreme learning machine (ELM) [12, 13] has emerged and attracted the attention from more and more researchers. ELM is a single-hidden layer feedforward network (SLFN) which randomly selects input weights and hidden neuron biases without training. The output weights are analytically determined by Moore-Penrose generalized inverse. The input weights are the weights of the connections between input neurons and hidden neurons and the output weights are the weights of the connections between hidden neurons and output neurons. ELM overcomes the proposed challenges faced by SVMs. Solving the regularized least square problem in ELM is faster than solving the quadratic programming problem in standard SVMs. Recent papers show that predicting accuracy achieved by ELM is comparable with or even higher than that of SVMs [14, 15]. Besides, ELM can learn the training data not only one-by-one but also chunk-by-chunk [16], so it provides a promising method to deal with big data. Moreover, traditional kernel methods can also be applied to ELM [15]. Due to its excellent features, ELM has been successfully used in many areas and there has been increasing research interest in it [17–19]. Many researchers have came up with some methods to improve ELM theories, such as ELMs for noisy/missing data [20, 21] and imbalanced data [22]. In [23], minimum class variance extreme learning machine (MCVELM) was proposed for human action recognition and achieved excellent performance. MCVELM borrows the idea from minimum class variance support vector machines (MCVSVM) [24] which is inspired from the optimization of Fisher’s discriminant ratio. The objective of MCVELM is to minimize both the intra-class variance of training data in the decision space and the training errors. However, the MCVELM does not take the structural risk into consideration, so it can lead to the over-fitting problem.

Inspired by its successful application to classification problems, in this paper, we introduce ELM into spoken language recognition, hoping to open up a new research direction for language recognition community. A new method referred to as regularized minimum class variance extreme learning machine (RMCVELM) is proposed and used as classifier for language recognition. Different from MCVELM, the proposed RMCVELM also minimizes the output weight norm directly which can effectively prevent the over-fitting problem. Therefore, the proposed RMCVELM can be considered as minimizing empirical risk, structural risk, and the intra-class variance of the training data in the decision space simultaneously. Since RMCVELM is based on structural risk minimization, it can be expected to provide better generalization ability.

To evaluate the effectiveness of the proposed RMCVELM, in this work, two types of acoustic features are used, GMM supervectors (GSVs) [25] and i-vectors [26]. The GSV has been widely used for many years and the i-vector approach, which provides an elegant way to capture the majority of useful variabilities of an utterance, has proven to be one of the most successful methods for language recognition nowadays [27, 28].

The outline of the paper is as follows. In Section 2, we introduce the ELM algorithm briefly. The proposed RMCVELM is described in detail in Section 3. In Section 4, experimental setup is introduced. In Section 5, we demonstrate the potential of the algorithm by applying it to language recognition task. In Section 6, computational complexity is analyzed. Finally, some conclusions are given in Section 7.

## 2 Overview of extreme learning machine

### 2.1 Basic extreme learning machine

ELM was first proposed by Huang [12] as an extremely simple and efficient method to train the single-hidden layer feedforward network. He has proven that the SLFN network with randomly chosen input weights and hidden neuron biases can approximate any continuous function to any desirable accuracy [29]. The input weights and hidden biases need not be tuned after randomly selected, and the output weights are analytically calculated using Moore-Penrose generalized pseudo-inverse.

Let us denote the training data as *ℵ*={(**x**
_{
i
},**t**
_{
i
})|**x**
_{
i
}∈**R**
^{
d
},*i*=1,…,*N*}, where *d* is the dimension of training data and *N* is the number of training data. **t**
_{
i
} is the label of **x**
_{
i
} which is a vector of length *m*, *m* is the number of classes. If **x**
_{
i
} belongs to class *j*, then all elements of **t**
_{
i
} are zero except the *j*th element, which takes the value 1.

*L*is the number of hidden nodes.

**w**

_{ j }and

*b*

_{ j }are the input weights and bias of the

*j*th hidden node, respectively. Hidden node parameters

**w**

_{ j }and

*b*

_{ j }remain fixed after randomly generated. β

_{ j }is the output weights of the

*j*th hidden node. The output of hidden layer with respect to training data

**x**

_{ i }is

**h**(

**x**

_{ i }),

*j*th hidden node. Many activation function can be used, such as sigmod, sine, and RBFs. The most widely used is the sigmod function.

From the perspective of ELM, **h**(**x**
_{
i
}) is a feature map which transforms the sample **x**
_{
i
} into ELM feature space.

**H**is defined as:

the *j*th column of **H** is the *j*th hidden node output with respect to inputs **x**
_{1},…,**x**
_{
N
}.

The orthogonal projection method can be efficiently used in ELM, namely **H**
^{
†
}=(**H**
^{
T
}
**H**)^{−1}
**H**
^{
T
} if **H**
^{
T
}
**H** is nonsingular or **H**
^{
†
}=**H**
^{
T
}(**H**
**H**
^{
T
})^{−1} if **H**
**H**
^{
T
} is nonsingular.

**x**is scored using the following equation,

### 2.2 Regularized extreme learning machine

*C*is the user specified parameter and provides a tradeoff between the training error and the regularization term \({\parallel \boldsymbol {\beta } {\parallel ^{2}_{2}}}\). Similar to ELM, the solution to RELM can be written as,

In [13], Huang has discussed the similarities between RELM and SVM and proven that to minimize the norm of the output weight ∥β∥_{2} is actually to maximize the distance of the separating margins of two different classes in the ELM feature space, which is similar to SVM. This conclusion is quite important and will be further discussed when introducing the RMCVELM.

## 3 Regularized minimum class variance extreme learning machine

Based on the previous introduction, in this section, we will describe RMCVELM in detail. We will first introduce the MCVELM and show some of it advantages when dealing with classification problems under certain conditions. Then the drawback of this algorithm is illustrated and RMCVELM is proposed to overcome the challenge faced by MCVELM.

### 3.1 Minimum class variance extreme learning machine

*t*

*r*() is the trace of a matrix and ξ

_{ i }=[

*ξ*

_{ i,1},…,

*ξ*

_{ i,m }]

^{ T }is the training error vector of the

*m*output nodes with respect to the training sample

**x**

_{ i }.

**S**

_{ w }is the within-class scatter matrix of the training samples in the ELM space defined by

*N*

_{ k }is the number of training samples in the class

*k*and \(\mathbf {m}_{k} = \frac {1}{N_{k}}\sum _{i=1}^{N_{k}}\mathbf {h}\left (\mathbf {x}_{i}\right)\) is the mean vector of class

*k*. So the term

*t*

*r*(β

^{ T }

**S**

_{ w }β) represents intra-class variance of all the training samples in the decision space. The solution to MCVELM can be written as

### 3.2 Regularized minimum class variance extreme learning machine

From another point of view, as stated in Section 2.2, regularization term \({\parallel \boldsymbol {\beta } {\parallel ^{2}_{2}}}\) can also be considered as structural risk. According to statistical learning theory, the real prediction risk of learning is consisted of empirical risk and structural risk. A model with good generalization ability should have the best trade-off between the two risks. However, MCVELM does not take the structural risk into consideration. Therefore MCVELM can not always achieve good generalization performance.

As stated before, \(\parallel \boldsymbol {\beta } {\parallel ^{2}_{2}}\) is the regularization term which aims at finding maximum separating margin in the ELM space and preventing over-fitting problem and *t*
*r*(β
^{
T
}
**S**
_{
w
}
β) represents intra-class variance of all the training samples in the decision space.

**x**is scored by the following equation,

The RMCVELM algorithm is summarized as follows.

We should note that RMCVELM can be considered as a unified mode for basic ELM, RELM, and MCVELM. When *C*
_{1}=*C*
_{2}=0, RMCVELM is equivalent to basic ELM. If *C*
_{1}=0 RMCVELM is equivalent to MCVELM and if *C*
_{2}=0 RMCVELM is equivalent to RELM. From this point of view, RMCVELM takes advantage of the good features of ELM, RELM, and MCVELM, so it has the potential to deal with more complicated data and tends to achieve more robust performance. In Figs. 2 and 3, we also show the hyperplanes obtained by RMCVELM. It is easy to note that the hyperplanes found by RMCVELM are between the hyperplanes of RELM and MCVELM. In Fig. 2, RMCVELM can get a hyperplane which is consistent with the data distribution. In Fig. 3, the hyperplanes found by RMCVELM not only reflect the data distribution but also use the samples near the margin which play an important role in classification. So RMCVELM can work well in these two cases.

## 4 Experimental setup

### 4.1 Training, development and evaluation data

The experimental setup for this work is based on the NIST 2009 Language Recognition Evaluation (LRE) [31]. The 2009 LRE consisted of 23 linguistic classes. The 23 linguistic classes are Amharic, Bosnian, Cantonese, Creole, Croatian, Dari, English-American, English-Indian, Farsi, French, Georgian, Hausa, Hindi, Korean, Mandarin, Pashto, Portuguese, Russian, Spanish, Turkish, Ukrainian, Urdu, and Vietnamese. Both conversational telephone speech utterances (CTS) and narrow band telephone segments from Voice of America broadcasts (VOA) are used as data sources for the 23 linguistic classes.

The training data consists of both CTS and VOA data. The CTS samples come from previous NIST LRE (2007) and the CallFriend, CallHome, OGI-22 collections. The VOA data consist of narrow band segments from VOA broadcasts. We totally have about 30,000 samples for the whole training set.

The calibration back-end described in Section 4.3 was trained on development dataset, which comprises data from the dataset provided by NIST for the 2003, 2005, and 2007 LRE and VOA. Only data of the 23 target languages are used. This set has about 10,000 samples.

The evaluation set consists of all the data from the NIST 2009 LRE. Experiments are performed on the 23 languages closed-set task. The criterion for evaluation is pooled equal error rate (EER) and average cost *C*
_{avg} defined by NIST.

### 4.2 Feature extraction

For feature extraction, we first extract 13-dimensional MFCC features and the cepstral features are processed with RASTA filtering. Then SDCC features [8] are used with a 7-1-3-7 parameterization.

A language and gender independent UBM [32] is trained using all of the training data with eight iterations of EM adapting all parameters-means, mixture weights, and diagonal covariances. The number of mixture components is 512.

After that, two different kinds of features are extracted for each utterance, namely GMM Supervectors (GSVs) [25] and i-vectors [26].

where *λ*
_{
i
}, Σ
_{
i
} and **m**
_{
i
} are the mixture weight, covariance and mean of the *i*th GMM component. A 28,672 dimensional GSV is extracted for each utterance. The GSVs are then compensated using Nuisance Attribute Projection (NAP) [33].

where **m** is the language- and channel-independent mean supervector (which can be taken to be the UBM supervector), **T** is a rectangular matrix of low rank and the i-vector **w** is a random vector having a standard normal distribution. For a given utterance, we use the standard method to extract the i-vector as described in [26]. The dimension of each i-vector is 400.

### 4.3 Calibration back-end

The calibration back-end [34] uses the scores from the classifiers described above rather than the GSVs or the i-vectors. It is trained on the separate development dataset described in Section 4.1 to obtain well-calibrated scores. Here, we use the MMI back-end [35] for calibration.

## 5 Experimental result

### 5.1 GSV systems

In order to evaluate the effectiveness of the proposed RMCVELM algorithm on language recognition performance, we first compare its performance with that of SVM, MCVELM, and RELM on the GSV features.

In our RMCVELM system, the input weight **w**
_{
i
} is randomly generated between −0.5 and + 0.5, and *b*
_{
i
} is randomly generated between 0 and + 1. The number of hidden node is 20,000. Here we use cross-validation technique on the training data to determine the parameters *C*
_{1} and *C*
_{2}. The training data is partitioned into four groups. Then three of the groups are used to train a set of models that are then evaluated on the remaining group. This procedure is then repeated for all four possible choices for the groups, and the four performance scores are then averaged. According to this cross-validation technique, the optimal parameters are set to be *C*
_{1}=2000 and *C*
_{2}=5.

In the SVM system, we use the proposed one-versus-rest strategy which is the standard method widely used for language recognition. The C value for SVM is set to be 0.8, it is tuned by the cross-validation technique as stated in the RMCVELM system.

In the MCVELM system, *C*
_{1} is set to be 0 and *C*
_{2}=1.5 which is achieved by the cross-validation technique stated above. Similarly, in the RELM system, *C*
_{2} is set to be 0 and *C*
_{1}=3000. In the Basic ELM system, *C*
_{1}=*C*
_{2}=0.

We also train a standard neural network (NN) with the same architecture as the ELM for comparison. The hidden node number is 120 and learning rate is 0.05. Besides, dropout method is used, the dropout fraction is 0.8. In addition, the weight penalty term based on *L*
^{2} norm is also taken into consideration. The coefficient of the weight penalty term is 0.005. We find that the dropout method and the weight penalty term must be used or there will be a serious over-fitting problem. All the above parameters are set by the cross-validation technique stated above.

Comparison of EERs (%) and *C*
_{avg} for different training methods based on GSV at different durations for the 23 proposed languages of LRE09

System | EER |
| ||||
---|---|---|---|---|---|---|

30 s | 10 s | 3 s | 30 s | 10 s | 3 s | |

SVM | 3.21 | 8.77 | 21.88 | 3.24 | 8.76 | 21.69 |

ELM | 4.32 | 12.17 | 25.93 | 4.30 | 12.46 | 26.13 |

MCVELM | 4.12 | 11.58 | 26.17 | 4.20 | 11.58 | 26.17 |

RELM | 2.89 | 7.79 | 22.00 | 2.82 | 7.86 | 21.97 |

NN | 2.97 | 7.98 | 22.56 | 2.95 | 7.97 | 22.56 |

RMCVELM | 2.70 | 7.59 | 20.74 | 2.62 | 7.60 | 20.62 |

First, we can easily note that the proposed RMCVELM outperforms SVM at all the three different durations. Especially, RMCVELM can greatly enhance the performance at the 30-s duration, obtaining at most 15 % (EER) and 19 % (*C*
_{avg}) relative improvement compared with SVM, respectively. In addition, RMCVELM works quite better than MCVELM and RELM, which verifies our theory. RMCVELM is more robust compared with MCVELM and RELM.

We should also note that the NN is better than the basic ELM which has no regularization term and comparable to the RELM. This is because the regularization term plays an important role in preventing the over-fitting problem. However, the NN does not achieve better performance than RELM, one reasonable explanation is that the NN tends to have the local minima problem. The advantages of ELM compared to NN has been discussed in detail in [12] which shows that ELM can achieve better performance than NN in many applications.

### 5.2 I-vector systems

For the i-vector inputs, all the i-vectors are centered on the origin and then linear discriminant analysis(LDA) is used to reduce the dimension. Since we have 23 target languages, each i-vector is reduced to 22 dimensions. Each i-vector is unit normalized after LDA, here we use **l**
_{
i
} to denote the normalized i-vector after LDA for a given utterance *i*.

**m**

_{ l }is the model of language

*l*,

**l**

_{test}is scored as:

where μ
_{
l
} is the mean vector for language *l* and Σ is the covariance matrix.

Then SVM, NN, and RMCVELM are applied to the i-vector. The C value for SVM is 1.2. For the NN, the hidden node number is 80, and learning rate is 0.01. The dropout fraction is 0.45, and the coefficient of the *L*
^{2} norm weight penalty term is 0.006. For the RMCVELE system, The number of hidden node is 3000 (experiments show that there is no significant improvement when the number of hidden node increases), and *C*
_{1}=2100,*C*
_{2}=3. All the parameters are determined as that in the GSV feature space.

Comparison of EERs (%) and *C*
_{avg} for different training methods based on i-vector at different durations for the 23 proposed languages of LRE09

System | EER |
| ||||
---|---|---|---|---|---|---|

30 s | 10 s | 3 s | 30 s | 10 s | 3 s | |

CDS | 2.63 | 6.72 | 20.22 | 2.59 | 6.80 | 20.36 |

GB | 2.65 | 6.77 | 19.90 | 2.60 | 6.86 | 20.02 |

SVM | 2.70 | 6.78 | 19.92 | 2.62 | 6.97 | 19.80 |

NN | 2.95 | 7.87 | 21.32 | 2.93 | 7.92 | 21.54 |

RMCVELM | 2.65 | 6.67 | 19.65 | 2.58 | 6.79 | 19.70 |

In addition, we can easily note that the CDS and GB which are used as baselines can achieve similar performance. Here we think that the CDS can approximate the GB. Since all i-vectors are unit vectors and centered on the origin, the mean of the Gaussian distribution for each language is approximately on the circle centered on the origin. The CDS averages i-vectors for a language, the model can roughly be regarded as a vector pointing from the origin to the mean of the Gaussian distribution for this language, so the separating margin between two classes is a linear hyperplane which is through the origin and equally divides the angle formed by the models of the two classes. The GB uses a shared covariance matrix, so the separating margin between two classes is also a linear hyperplane. If the covariance matrix is a diagonal matrix and the diagonal elements are the same, the GB hyperplane also will be through the origin and will be quite close to the CDS hyperplane. In this experiment, we find that the main diagonal elements of the full covariance matrix are much larger than other elements, so the full covariance matrix is close to a diagonal matrix. In addition, the main diagonal elements are close to each other. In this case, we can assume that the CDS is almost equivalent to the GB. So the two methods can achieve similar performance.

Besides, the NN can not get good performance as it is often very sensitive to many parameters, such as learning rate and dropout fraction, and tends to have the local minima problem.

### 5.3 Discussion

The experimental results show that in most cases, RMCVELM can achieve better performance than SVM. Compared with SVM, RMCVELM has some potential advantages.

In [13], the differences and similarities between SVM and ELM has been discussed. ELM and SVM have similar optimization formula. However, in SVM optimal Lagrange multipliers *α*
_{
i
} are found from the hyperplane \( \sum _{i=1}^{N} \alpha _{i} t_{i} = 0\), so SVM often obtains a suboptimal optimization. Different from SVM, ELM does not need to satisfy this condition, so ELM has milder optimization constraints and can find *α*
_{
i
} from the entire space. RMCVELM, which is an extension of ELM, naturally inherits this good feature.

In addition, RMCVELM can be considered as a unified mode for ELM, regularized ELM, and MCVELM. It considers the training error, generalization ability, and distribution of all the training data in the ELM space simultaneously. By adjusting parameters *C*
_{1} and *C*
_{2}, RMCVELM can learn much better even when the data distribution varies greatly. So it has the potential to deal with more complicated data. From this point of view, RMCVELM is more robust.

## 6 Computational complexity

**H**

^{ T }

**H**+

*C*

_{1}

**I**+

*C*

_{2}

**S**

_{ w }) is

*L*∗

*L*. In most applications, the number of hidden node

*L*is much smaller than the number of training data:

*L*<<

*N*. So RMCVELM can greatly reduce the training time.

## 7 Conclusions

In this paper, we proposed a novel algorithm regularized minimum class variance extreme learning machine for language recognition. The RMCVELM aims at minimizing empirical risk, structural risk, and the intraclass variance of the training data in the decision space simultaneously. The proposed method improves the performance for language recognition and can learn at a much faster speed compared with traditional SVM. In addition, RMCVELM can also be applied to i-vector space and achieves an equivalent performance compared with the CDS and GB method. To our best knowledge, this is the first time such an approach is applied to language recognition.

## Declarations

### Acknowledgements

This work is supported by National Natural Science Foundation of China (Project 61370034, Project 61273268 and Project 61005019).

**Open Access** This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

## Authors’ Affiliations

## References

- H Li, B Ma, KA Lee, Spoken language recognition: from fundamentals to practice. Proc. IEEE. 101(5), 1136–1159 (2013).View ArticleGoogle Scholar
- F Biadsy, Automatic dialect and accent recognition and its application to speech recognition, PhD thesis, Columbia University (2011).Google Scholar
- MA Zissman, KM Berkling, Automatic language identification. Speech Comm. 35(1), 115–124 (2001).MATHView ArticleGoogle Scholar
- YK Muthusamy, E Barnard, RA Cole, Reviewing automatic language identification. Signal Proc. Mag. IEEE. 11(4), 33–41 (1994).View ArticleGoogle Scholar
- MA Zissman, Comparison of four approaches to automatic language identification of telephone speech. IEEE Transactions on Speech and Audio Processing. 4(1), 31 (1996).View ArticleGoogle Scholar
- WM Campbell, FS Richardson, DA Reynolds, in
*ICASSP (4)*. Language recognition with word lattices and support vector machines (IEEE,Honolulu, USA, 2007), pp. 989–992.Google Scholar - W-Q Zhang, W-W Liu, Z-Y Li, Y-Z Shi, J Liu, Spoken language recognition based on gap-weighted subsequence kernels. Speech Communication. 60, 1–12 (2014).View ArticleGoogle Scholar
- PA Torres-Carrasquillo, E Singer, MA Kohler, RJ Greene, DA Reynolds, JR Deller Jr, in
*INTERSPEECH*. Approaches to language identification using gaussian mixture models and shifted delta cepstral features (ISCA,Denver, USA, 2002).Google Scholar - W-Q Zhang, L He, Y Deng, J Liu, MT Johnson, Time-frequency cepstral features and heteroscedastic linear discriminant analysis for language recognition. IEEE Trans. Audio, Speech, and Lang. Process. 19(2), 266–276 (2011).View ArticleGoogle Scholar
- WM Campbell, E Singer, PA Torres-Carrasquillo, DA Reynolds, in
*ODYSSEY04-The Speaker and Language Recognition Workshop*. Language recognition with support vector machines (ISCA,Toledo, Spain, 2004).Google Scholar - WM Campbell, JP Campbell, DA Reynolds, E Singer, PA Torres-Carrasquillo, Support vector machines for speaker and language recognition. Comput. Speech & Lang. 20(2), 210–229 (2006).View ArticleGoogle Scholar
- G-B Huang, Q-Y Zhu, C-K Siew, in
*Neural Networks, 2004. Proceedings. 2004 IEEE International Joint Conference On*, 2. Extreme learning machine: a new learning scheme of feedforward neural networks (IEEE,Budapest, Hungary, 2004), pp. 985–990.Google Scholar - G-B Huang, DH Wang, Y Lan, Extreme learning machines: a survey. Int. J. Mach. Mach. Learn. Cybern. 2(2), 107–122 (2011).View ArticleGoogle Scholar
- G-B Huang, Q-Y Zhu, C-K Siew, Extreme learning machine: theory and applications. Neurocomputing. 70(1), 489–501 (2006).View ArticleGoogle Scholar
- G-B Huang, H Zhou, X Ding, R Zhang, Extreme learning machine for regression and multiclass classification. Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on. 42(2), 513–529 (2012).View ArticleGoogle Scholar
- N-Y Liang, G-B Huang, P Saratchandran, N Sundararajan, A fast and accurate online sequential learning algorithm for feedforward networks. Neural Netw. IEEE Trans. 17(6), 1411–1423 (2006).View ArticleGoogle Scholar
- J Xu, H Zhou, G-B Huang, in
*Information Fusion (FUSION), 2012 15th International Conference On*. Extreme learning machine based fast object recognition (IEEE,Singapore, 2012), pp. 1490–1496.Google Scholar - MM Sole, M Tsoeu, in
*AFRICON, 2011*. Sign language recognition using the extreme learning machine (IEEE,Livingstone, Zambia, 2011), pp. 1–6.Google Scholar - S Suresh, V Babu, N Sundararajan, in
*Control, Automation, Robotics and Vision, 2006. ICARCV’06. 9th International Conference On*. Image quality measurement using sparse extreme learning machine classifier (IEEE,Singapore, 2006), pp. 1–6.View ArticleGoogle Scholar - P Horata, S Chiewchanwattana, K Sunat, Robust extreme learning machine. Neurocomputing. 102, 31–44 (2013).View ArticleGoogle Scholar
- Q Yu, Y Miche, E Eirola, M Van Heeswijk, E Séverin, A Lendasse, Regularized extreme learning machine for regression with missing data. Neurocomputing. 102, 45–51 (2013).View ArticleGoogle Scholar
- W Zong, G-B Huang, Y Chen, Weighted extreme learning machine for imbalance learning. Neurocomputing. 101, 229–242 (2013).View ArticleGoogle Scholar
- A Iosifidis, A Tefas, I Pitas, Minimum class variance extreme learning machine for human action recognition (2013).Google Scholar
- S Zafeiriou, A Tefas, I Pitas, Minimum class variance support vector machines. Image Process. IEEE Trans. 16(10), 2551–2564 (2007).MathSciNetView ArticleGoogle Scholar
- WM Campbell, DE Sturim, DA Reynolds, Support vector machines using gmm supervectors for speaker verification. Signal Proc. Lett. IEEE. 13(5), 308–311 (2006).View ArticleGoogle Scholar
- N Dehak, P Kenny, R Dehak, P Dumouchel, P Ouellet, Front-end factor analysis for speaker verification. Audio, Speech, and Lang. Process, IEEE Trans. 19(4), 788–798 (2011).View ArticleGoogle Scholar
- N Dehak, PA Torres-Carrasquillo, D Reynolds, R Dehak, in
*INTERSPEECH 2011*. Language Recognition via I-Vectors and Dimensionality Reduction (ISCA,Florence, Italy, 2011), pp. 857–860.Google Scholar - D Martınez, O Plchot, L Burget, O Glembek, P Matejka, in Proceedings of Interspeech. Language recognition in ivectors space (Firenze, Italy, 2011), pp. 861–864.Google Scholar
- G-B Huang, Learning capability and storage capacity of two-hidden-layer feedforward networks. Neural Networks, IEEE Transactions on. 14(2), 274–281 (2003).View ArticleGoogle Scholar
- CM Bishop,
*Pattern Recognition and Machine Learning*, vol. 1 (Springer, New York, 2006).MATHGoogle Scholar - AF Martin, CS Greenberg, in
*Odyssey*. The 2009 nist language recognition evaluation (ISCA,Brno, Czech, 2010), p. 30.Google Scholar - DA Reynolds, TF Quatieri, RB Dunn, Speaker verification using adapted gaussian mixture models. Digital signal processing. 10(1), 19–41 (2000).View ArticleGoogle Scholar
- WM Campbell, DE Sturim, DA Reynolds, A Solomonoff, in
*Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings. 2006 IEEE International Conference On*, 1. Svm based speaker verification using a gmm supervector kernel and nap variability compensation (IEEE,Toulouse, France, 2006).Google Scholar - M Penagarikano, A Varona, M Diez, LJ Rodriguez-Fuentes, G Bordel, in INTERSPEECH. Study of different backends in a state-of-the-art language recognition system, (2012).Google Scholar
- W ZHANG, T HOU, J LIU, Discriminative score fusion for language identification. Chin. J. Electron. 19(1), 124–128 (2010).Google Scholar
- E Singer, P Torres-Carrasquillo, D Reynolds, A McCree, F Richardson, N Dehak, D Sturim, in
*Acoustics, Speech and Signal Processing, 2007. ICASSP 2007. IEEE International Conference On*. The mitll nist lre 2011 language recognition system (ISCA,Singapore, 2012), pp. 209–215.Google Scholar