Research Article  Open Access
ZhiXia Yang, YuanHai Shao, YaoLin Jiang, "A Learning Framework of Nonparallel Hyperplanes Classifier", The Scientific World Journal, vol. 2015, Article ID 497617, 12 pages, 2015. https://doi.org/10.1155/2015/497617
A Learning Framework of Nonparallel Hyperplanes Classifier
Abstract
A novel learning framework of nonparallel hyperplanes support vector machines (NPSVMs) is proposed for binary classification and multiclass classification. This framework not only includes twin SVM (TWSVM) and its many deformation versions but also extends them into multiclass classification problem when different parameters or loss functions are chosen. Concretely, we discuss the linear and nonlinear cases of the framework, in which we select the hinge loss function as example. Moreover, we also give the primal problems of several extension versions of TWSVM’s deformation versions. It is worth mentioning that, in the decision function, the Euclidean distance is replaced by the absolute value , which keeps the consistency between the decision function and the optimization problem and reduces the computational cost particularly when the kernel function is introduced. The numerical experiments on several artificial and benchmark datasets indicate that our framework is not only fast but also shows good generalization.
1. Introduction
Classification problem is an important issue in machine learning and data mining, which is mainly comprised of binary and multiclass classification. Support vector machine (SVM), proposed by Burges [1] and Cortes and Vapnik [2], is an excellent tool for classification. In contrast with conventional artificial neural networks (ANNS) which aim at reducing empirical risk, SVM is principled and implements the structural risk minimization (SRM) that minimizes the upper bound of the generalization error [3–5]. Within a few years after its introduction, SVM has been successfully applied to pattern classification and regression estimation like face detection [6, 7], text categorization [8], time series prediction [9], bioinformatics [10], and so forth.
Recently, for binary classification, Mangasarian and Wild [11] proposed the generalized eigenvalue proximal support vector machine (GEPSVM) via two nonparallel hyperplanes. In their approach, the data points of each class are proximal to one of two nonparallel hyperplanes. The nonparallel hyperplanes are determined by eigenvectors corresponding to the smallest eigenvalues of two related generalized eigenvalue problems. Inspired by GEPSVM [11], Jayadeva et al. [12] developed twin SVM (TWSVM) with two nonparallel hyperplanes. However, the two hyperplanes are got by solving two quadratic programming (QP) problems, similar to the standard SVM. Furthermore, TWSVM differs from the standard SVM in fundamental way. In TWSVM, one solves a pair of smaller size QP problems rather than a single QP problem in the standard SVM. Therefore, TWSVM works faster than the standard SVM. Subsequently, there are many extensions for TWSVM including the improvements on TWSVM (TBSVM) [13], the least square TWSVM (LSTWSVM) [14–17], nonparallel plane proximal classifier (NPPC) [18], smooth TWSVM [19], geometric algorithm [20], and twin support vector regression (TWSVR) [21]. TWSVM was also extended to deal with multiclassification TWSVM [22–24]. More precisely, in [22], TWSVM was extended straight from binary classification to multiclass classification, in which each primal problem covers all patterns except the patterns of the th class in the constraints for the th () hyperplane. In [23], the authors extended TWSVM based on the idea of “oneversusrest” (1vr) from binary classification to multiclass classification, in which there are two quadratic programming (QP) problems for each reconstructing binary classification. However, they both have not kept the advantage of TWSVM which has lower computational complexity than that of the standard SVM. In [24], Yang et al. proposed multiple birth SVM (MBSVM) with much lower computational complexity than that of both [22, 23] by solving smaller size of QP problems for class classification; only the empirical risk is considered like TWSVM. However, in TBSVM [13], the structural risk minimization principle is implemented by introducing the regularization term.
In this paper, we propose a novel learning framework of nonparallel hyperplanes support vector machines based on TWSVM and its extension versions, called NPSVMs, which not only provide a unified view for TWSVM and its many extension versions but also can deal with binary and multiclass classification problems. For binary classification, if the loss function is the hinge loss function, then the framework can become TWSVM [12] or TBSVM [13] with different parameters; if the loss function is the square loss function, then the framework is LSTWSVM [14]; if the loss function is the convex combination of the linear and square loss functions, then the framework is NPPC [18]. Actually, we can also get smooth TWSVM [19] by replacing 2norm with 1norm in the framework. However, for multiclass classification, the framework does not directly extend, in which we switch the roles of the patterns of the th class and the rest class and replace “min” with “max” in the decision function. Moreover, we only use the absolute value rather than the Euclidean distance in the decision function due to the twofold reasons: reducing the computational cost particularly when the kernel function is introduced and making the consistency since it is the corresponding absolute value that appears in the primal problems. Concretely, we discuss the linear and nonlinear cases of the framework, in which we select the hinge loss function as example. Moreover, we also give the primal problems of extensions of LSTWSVM, 1norm LSTWSVM, NPPC, and smooth TWSVM. Finally, the numerical experiments on several artificial and benchmark datasets indicate that our frameworks are not only fast but also show good generalization.
The paper is organized as follows. Section 2 introduces the brief reviews of SVMs. Section 3 proposes our frameworks, in which Section 3.1 discusses the linear framework, Section 3.2 extend into the nonlinear framework, Section 3.3 gives SOR algorithm for solving the hinge NPSVMs, and Section 3.4 discusses several other extension approaches. Finally, Section 4 deals with experimental results and Section 5 contains concluding remarks.
2. Brief Reviews of SVMs
2.1. Twin Support Vector Machine
Given the following training set for the binary classification:where is the th data point, the input is a pattern, the output is a class label, , and is the number of data points. In addition, let and be the number of data points in positive class and negative class, respectively, and . Furthermore, the matrices and consist of the inputs of Class and the inputs of Class 2, respectively.
The gaol of TWSVM [12] is to find two nonparallel hyperplanes in dimensional input space:such that one hyperplane is close to the patterns of one class and far away from the patterns of the other class to some extent. TWSVM is in spirit of GEPSVM [11]. But both of GEPSVM and TWSVM are different from the standard SVM. For TWSVM, each hyperplane is generated by solving a QP problem looking like the primal problem of the standard SVM. The primal problems of TWSVM can be presented as follows:where and are nonnegative parameters and and are vectors of ones of appropriate dimensions. In the QP problem (4), the objective function tends to keep hyperplane (2) close to the patterns of Class 1 and the constraints require the hyperplane (2) to be at a distance of at least 1 from the patterns of Class 2. The QP problem (5) has similar property. Moreover, we note that the constraints do not contain all patterns in the training set (1) but are determined by only the patterns of one class in both classes. Therefore, in [12], the authors claimed that TWSVM is approximately four times faster than the standard SVM.
Define and . It has been shown that when both and are positive definites, the Wolfe duals of (4) and (5) are written as follows:respectively, where and are Lagrangian multipliers.
In order to avoid the possible illconditioning of and , TWSVM introduces a term (), where is an identity matrix of appropriate dimensions. Thus, the nonparallel hyperplanes (2) and (3) can be obtained from the solutions and of the QP problems (6) and (7). Considerwhere , . Moreover, a new pattern is assigned to Class (), depending on which of the two nonparallel hyperplanes given by (2) and (3) lies closer to; that is,
2.2. Multiple Birth Support Vector Machine
Given the training setwhere the input , , is the pattern and the output is the class label. The task is to seek hyperplanes,and assign the class label according to which hyperplane a new pattern is farthest from.
For convenience, denote the number of data points of the th class in the training set (10) as and define the following matrixes: the patterns belonging to the th class are represented by the matrix , . In addition, define the matrixthat is, consists of the patterns belonging to all classes except the th class, . The primal problems of MPSVM [24] are comprised of the following QP problem:where and are the vectors of ones, is the slack variable, and is the penalty parameter, . The dual problem of QP problem (13) is formulated as follows:where the penalty parameter , , and , . Similarly, in order to avoid the possibility of the illconditioning of the matrix in some situations, one introduces a regularization term , where is a fixed small scalar and is the identity matrix with appropriate size.
After getting the solution to the above QP problem (13) with , a new pattern is assigned to class (), depending on which of the hyperplanes given by (11) lies farthest from; that is, the decision function is represented aswhere is the absolute value.
3. The Framework of Nonparallel Hyperplanes Classifiers
In this section, we propose a learning framework of nonparallel hyperplanes classifier, which gives a unified form for TWSVM and its many extension versions and extend them into multiclass classification problem. We first develop the linear framework and then extend it to nonlinear framework.
3.1. Linear Framework
Given the training set (10), the task is to find nonparallel hyperplanes:one for each class. For obtaining the unknown hyperplanes, we construct the following standard framework for each unknown hyperplane:where the matrix is comprised of the patterns in the th class, the matrix is defined (12), and are the parameters, and are vectors of ones of appropriate dimensions, , and is the loss function (e.g., square loss, hinge loss, etc.). In the optimization problem (17), the first term approximatively minimizes the sum of the squared Euclidean distances from the patterns except for the th class to hyperplanes; the second term is the Tikhonov regularization term [25] and can implement the structural risk minimization principle like TBSVM [13]; the third term constitutes the loss function which is defined different loss functions corresponding to different models.
For a new pattern , we assign to class () according to the following decision function:where is the absolute value. Note that we only use the absolute value in the decision function. There are two main reasons: one is that the first term of the optimization problem (17) just minimizes the sum of the square rather than the sum of square Euclidean distance from the patterns to hyperplanes, so it should keep consistency between the optimization problem and the decision function; another is that it reduces the computational cost particularly when the kernel function is introduced afterwards.
In fact, if , the parameter is equal to 0, and the loss function is hinge loss function, that is, , then the optimization problem (17) becomes TWSVM [12]. Moreover, if the parameter is alterable, then it is TBSVM [13]. And if the loss function is the square loss function, that is, , it is LSTWSVM [14]. And if the loss function is a convex combination of linear and square loss, that is, , where , then it is NPPC [18]. Other extension versions of TWSVM also can be contained in the optimization problem (17), for instance, smooth TWSVM, 1norm LSTWSVM [17], and so forth, in which we just need to select proper norm or loss function.
More importantly, our framework can solve multiclass classification problem, which is extension of TWSVM, TBSVM, LSTWSVM, NPPC, and so forth. It should be pointed out that our framework is not straight extension of TWSVM and its deformation versions. Concretely, from the optimization problem (17), we can see that the first term contains the patterns except for those of the th class and the third term just involves the patterns of the th class. This strategy cannot lead to significant increase of the complexity of the optimization when the number of classes increases. We will dwell on in specific algorithm afterwards.
Now, we give the detailed algorithm to the hinge loss function as an example, called hinge NPSVM (HNPSVM). And then the optimization problem (17) is the following formulation with the hinge loss function:where the matrix is comprised of the patterns in the th class, the matrix is defined (12), and are the parameters, and are vectors of ones of appropriate dimensions, and . Actually, the problem is equivalent to the following quadratic programming:where the matrix is comprised of the patterns in the th class, the matrix is defined (12), and are the parameters, and are vectors of ones of appropriate dimensions, and .
In fact, for , we have QP problems like (20). In particular, when is equal to 2, that is, , the QP problems (4) and (5) can be obtained as a special case of (20) with . For simplicity, assume that the number of each class points is almost balanced; namely, the number of the th class is . Then, note that the constraints just involve the patterns of the th class, so the complexity of the the problem (20) is no more than . However, if TWSVM is directly extended to multiclass classification case like [22], we will get a different optimization problem, in which the roles of patterns of the th class and the rest class are switched. Thus, the complexity of the optimization problem will increase significantly and is determined by the patterns except for the patterns of the th class in the training set (10), which is no more than . Obviously, our approach is approximately times faster than the model in [22]. On the other hand, when the number of each class points is unbalanced, our apprach still is faster than the model in [22] because the complexity of our optimization problem just is decided by the number of the patterns of the th class rather than the patterns of the rest classes. Therefore, our HNPSVM keeps the computation complexity low.
It is well known that the solution of primal problem (20) is obtained from the solutions of their dual problems. So we now derive their dual problems. The Lagrangian function of the problem (20) is given bywhere , are nonnegative Lagrange multiplier vectors. The KarushKuhnTucker (KKT) necessary and sufficient optimality conditions [26] for the QP problem (20) are given bySince , according to (24), we haveNext, from (22) and (23), we can obtainwhere is an identity matrix of appropriate dimensions. Let ; (29) can be written aswhere and . And then putting (30) into the Lagrangian function (21) and using (22)–(28), we can get the dual problem of the primal problem (20):where and are parameters and . Obviously, if we have the solution of the QP problem (31), then we obtain the nonparallel hyperplanes (16) by (30).
It is worth mentioning that the parameter replaces as in (8), so is no longer a fixed small scalar but a weighting factor which determines the tradeoff between the regularization term and the empirical risk in the problem (20). Therefore, the high and low of the value of reflects the structure of minimization principle and our HNPSVM includes MBSVM.
3.2. Nonlinear Framework
Similarly, we also extend the linear framework of NPSVMs to nonlinear case. For a class classification (10), our goal is to find kernelgenerated hyperplanes:where and is an appropriately chosen kernel function.
In order to obtain the hyperlanes (32), we construct the following framework formulation:where the matrix is comprised of the patterns in the th class, the matrix is defined (12), and are the parameters, and are vectors of ones of appropriate dimensions, , and is the loss function (e.g., square loss or hinge loss, etc.). Similarly, as discussed in the last subsection, the problem (33) can be reduced to the nonlinear formulations of the difference approaches (e.g., TWSVM, TBSVM, LSTWSVM, NPPC, etc.) when the difference loss functions or parameters are selected for .
A new pattern is assigned to the th class by the following decision functions:where is the absolute value. Note that, in this decision function (34), we just compute the absolute value rather than Euclidean distance from the pattern to the hyperplanes. This strategy reduces the complexity of computation because Euclidean distance should be from the pattern to the th hyperplanes. Thus, the decision function (34) not only saves the computation quantity but also keeps the consistency with the first term of the problem (33).
Now, we still select the hinge loss function as example. Then, the problem (33) can be formulated as follows:where the matrix is comprised of the patterns in the th class, the matrix is defined by (12), and are parameters, and are vectors of ones of appropriate dimensions, and . Similarly, derived process with the linear case, its dual problem is formulated as:where and are parameters, , , and . And the augmented vector is given by .
3.3. SOR Algorithm
In our HNPSVMs, the QP problems (31) and (36) can be rewritten as the following unified forms:where is positive definite. For example, the above problem becomes the problem (36), when , .
The above problem (37) can be solved efficiently by the following successive overrelaxation (SOR) algorithm; see [27].
Algorithm 1. SOR for the QP problem (36) is as follows.(1)Select the parameter and the initial value .(2)Suppose that is obtained by the times iterate; compute according to the following iterate formula:where . And define , where and are the strictly lower triangular matrix and the diagonal matrix, respectively.(3)Stop if is less than some desired tolerance. Else, replace by and by and go to 2.
SOR is an excellent TWSVM solver, because it can process efficiently very large datasets that need not reside in memory. Furthermore, it has been proved that this algorithm converges linearly to a solution in [27, 28]. It should be pointed out that we employ the ShermanMorrisonWoodbury formula [29] for the inversion of matrix and, hence, need only to invert matrix with a lower order , instead of the order . Further, in practise, if the number of patterns in the th classe is large, then the rectangular kernel technique [30, 31] can be applied to reduce the dimensionality of our nonlinear classifiers.
3.4. Several Others Approaches
In this section, we briefly give several extension versions based on our framework by selecting different loss function or replacing 2norm.
First, if the square loss function is chosen, that is, , then we can get the following formulation from the framework (17):where the matrix is comprised of the patterns in the th class and the matrix is defined by (12), and are parameters, and are vectors of ones of appropriate dimensions, and . This is extension version of LSTWSVM [14].
Second, if we replace 2norm with 1norm in the problem (39), then we can get the extension of 1norm LSTWSVM [17] as follows:where the matrix is comprised of the patterns in the th class, the matrix is defined by (12), and are parameters, and are vectors of ones of appropriate dimensions, and .
Third, if the loss function is a convex combination of linear and square loss, that is, , where , then we can obtain extension version of NPPC [18] as follows:where the matrix is comprised of the patterns in the th class, the matrix is defined by (12), and are parameters, and are vectors of ones of appropriate dimensions, and .
Forth, if the square hinge loss function is selected, that is, , then we can get the extension version of smooth TWSVM as follows:where the matrix is comprised of the patterns in the th class, the matrix is defined by (12), and are parameters, and are vectors of ones of appropriate dimensions, and .
These approaches have the same decision function (18) and can be extended into nonlinear case. And their solving methods can construct based on their binary algorithms.
4. Numerical Experiments
In this section, we present experimental results of our binary HNPSVM (BHNPSVM) and multiclass HNPSVM (MHNPSVM) on both artificial and benchmark datasets. In experiments, we focus on the comparison between our methods and some stateoftheart classification methods, including SVM, GEPSVM, TWSVM, “1v1,” “1vr,” and MBSVM. All the classification methods are implemented in MATLAB 7.0 [32] environment on a PC with Intel P4 processor (2.9 GHz) with 1 GB RAM. In order to give the fastest training speed, we employ Libsvm [33] to implement the SVM, “1v1,” and “1vr”. Our BHNPSVM and MHNPSVM and TWSVM and MBSVM are implemented using SOR technique; GEPSVM is implemented by simple MATLAB functions like “eig,” respectively. As for the problem of selecting parameters, we employ standard 10fold crossvalidation technique [34]. Furthermore, the parameters for all methods are selected from the set .
4.1. Toy Examples
Firstly, we consider a simple twodimensional “Cross Planes” dataset as Example 1, which was tested in [11, 13] to indicate that nonparallel hyperplanes classifiers can handle the cross planes dataset much better compared with parallel ones. Now, we show that our BHNPSVM also can handle crossplanes type data well due to use of our decision function. The “Cross Planes” dataset is generated by perturbing points lying on two intersecting lines. Figures 1(a)–1(d) show the dataset and the linear classifiers obtained by SVM, GEPSVM, TWSVM, and our BNPSVM. It is easy to see that the result of our BNPSVM is more reasonable than that of SVM, and better than that of GEPSVM and TWSVM. In addition, we list the accuracy and CPU time for these four classifiers in Table 1. From Table 1, we can see that our BNPSVM obtains the best accuracy while not the slowest computing time.

(a) SVM
(b) GEPSVM
(c) TWSVM
(d) BNPSVM
Secondly, we consider a twodimensional threeclass dataset as Example 2 to show the operating mechanism of our MNPSVM and other multipleclass classifiers. The threeclass dataset is generated by perturbing points lying on three intersecting lines. Figures 2(a)–2(d) show the dataset and the linear classifiers obtained by “1v1,” “1vr,” MBSVM, and MHNPSVM. It is easy to see that the result of MBSVM and MHNPSVM is more reasonable than that of “1v1” and “1vr.” We also list the accuracy and CPU time of Example 2 for these four classifiers in Table 1. From Table 1, we can see that our MHNPSVM obtains the best accuracy in all these two examples, indicating that our MHNPSVM is suitable for both “Cross Planes” and multiclass problems.
(a) “1v1”
(b) MBSVM
(c) “1vr"
(d) MHNPSVM
4.2. Benchmark Datasets
In order to further compare our methods with others, we examine nine binaryclass datasets and nine multiclass datasets used by [12, 35], from the UCI Repository of machine learning database [36]. Table 2 gives the details of these eighteen datasets.
 
#Ins is the number of the training points; #attributes is the number of attributes; #class is the number of class. 
In order to compare the behavior of our linear BHNPSVM with SVM, GEPSVM, and TWSVM, the numerical experimental results for binaryclass UCI datasets are summarized in Table 3. In Table 3, the classification accuracy and computation time are listed. In Table 3, the best accuracy is shown by bold figures. It is easy to see that most of the accuracies of our linear BHNPSVM are better than linear SVM, GEPSVM, and TWSVM on these datasets. It can also be seen that our BHNPSVM is a little faster than TWSVM and is competitive with SVM (implements by Libsvm). We also list the mean accuracy and mean time for these four classifiers. Our BHNPSVM gains the the highest mean accuracy while faster training speed than TWSVM.
 
A greater difference between BHNPSVM and TWSVM. 
Table 4 is concerned with our kernel BHNPSVM, SVM, GEPSVM, and TWSVM on binaryclass UCI datasets. The Gaussian kernel is used. The kernel parameter is also obtained through searching from the range from to . The training CPU times for these four classifiers are also listed. The results in Table 4 are similar to those appearing in Table 3 and therefore confirm the above conclusion further.
 
A greater difference between BHNPSVM and TWSVM. 
In order to compare the behavior of our MHNPSVM with other multipleclass classifiers, we compare our MHNPSVM with “1v1,” “1vr,” and MBSVM, the linear results of numerical experiments on multiclass UCI datasets are summarized in Table 5. In Table 5, the classification accuracy and computation time are listed.
 
A greater difference between MHNPSVM and MBSVM. 
From Table 5, we can see that the accuracy of linear MHNPSVM is significantly better than linear MBSVM on all 9 UCI datasets. We also obtain that MHNPSVM and MBSVM are almost same fast because they both solve two SOR algorithms instead of two QP problems with the same size. In contrast, classification accuracy of “1v1” and “1vr” is no statistical difference with MHNPSVM for all cases except for vowel dataset, and “1v1” and “1vr” are a bit lower than MHNPSVM and MBSVM in average training time. Thus, with the proposed formulation of MHNPSVM allows the classifier to learn better by reducing the generalization errors. However, this improved performance is obtained at the cost of more tuning effort involved. This is because MHNPSVM requires tuning of more parameters than MBSVM.
Table 6 shows the nonlinear MHNPSVM with “1v1,” “1vr,” and MBSVM, the results of numerical experiments. In Table 6, the classification accuracy and computation time are listed. The results in Table 6 are similar to those appearing in Table 5; MHNPSVM has better classification accuracy than MBSVM in eight datasets, while MBSVM is better than MHNPSVM in one dataset, and MHNPSVM and MBSVM are much faster than “1v1” and “1vr”, especially when the amount of data increases.
 
A greater difference between MHNPSVM and MBSVM. 
5. Conclusions
In this paper, a general framework of nonparallel hyperplanes support vector machines, termed NPSVMs, are proposed for binary classification and multiclass classification. For binary classification, this framework includes TWSVM and its many deformation versions, for instance, TWSVM, TBSVM, LSTWSVM, NPPC, and so forth, when different loss functions and parameters are selected. For multiclass classification, we do not directly extend TWSVM and its deformation versions to get the framework, in which we switch the roles of the patterns of the th class and the rest classes. This strategy does not lead to significant increase of the computation complexity when the number of classes is increasing. Moreover, in the decision function, “” and Euclidean distance in TWSVM are replaced by “” and the absolute value , respectively. The absolute value is not only simpler but also more consistent with the primal problems. In particular, we discuss the linear and nonlinear case of the framework with the hinge loss function as example. Moreover, we also give the primal problems of several extensions of TWSVM’s deformation versions. The numerical experiments on several artificial and benchmark datasets indicate that our NPSVMs yield comparable generalization performance compared with SVM, GEPSVM, TWSVM, MBSVM, “1v1,” and “1vr”. In short, the proposed framework not only includes TWSVM and its many deformation versions but also extends them into multiclass classification under keeping the merit of TWSVM (learning speed).
In the future, we will develop the idea of nonparallel hyperplanes classifiers to other problems such as ordinal regression, multiinstance, and multilabel classification.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This work is supported by the National Natural Science Foundation of China (no. 11161045); the Project is sponsored by SRF for ROCS, SEM, and the Open Funding Project of the National Key Laboratory of Biochemical Engineering (no. 2013KF01) and the Zhejiang Provincial Natural Science Foundation of China (no. Q12A010077).
References
 C. J. C. Burges, “A tutorial on support vector machines for pattern recognition,” Data Mining and Knowledge Discovery, vol. 2, no. 2, pp. 121–167, 1998. View at: Publisher Site  Google Scholar
 C. Cortes and V. N. Vapnik, “Supportvector networks,” Machine Learning, vol. 20, no. 3, pp. 273–297, 1995. View at: Publisher Site  Google Scholar
 C. Nello and S. T. John, An Introduction to Support Vector Machines and Other KernelBased Learning Methods, Cambridge University Press, Cambridge, Mass, USA, 1st edition, 2000.
 N. Deng, Y. Tian, and C. Zhang, Support Vector Machines: Theory, Algorithms and Extensions, Chapman & Hall/CRC Data Mining and Knowledge Discovery Series, CRC press, 2013. View at: MathSciNet
 V. N. Vapnik, The Nature of Statistical Learning Theory, Springer, New York, NY, USA, 1995. View at: MathSciNet
 E. Osuna, R. Freund, and F. Girosi, “Training support vector machines: an application to face detection,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 130–136, IEEE, San Juan, Puerto Rico, June 1997. View at: Publisher Site  Google Scholar
 H. Cevikalp, B. Triggs, and V. Franc, “Face and landmark detection by using cascade of classifiers,” in Proceedings of the 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG '13), pp. 1–7, Shanghai, China, April 2013. View at: Publisher Site  Google Scholar
 T. Joachims, “Text categorization with support vector machines: learning with many relevant features,” in Machine Learning: ECML98, pp. 137–142, 1998. View at: Google Scholar
 L. J. Cao, “Support vector machines experts for time series forecasting,” Neurocomputing, vol. 51, no. 1–4, pp. 321–339, 2003. View at: Publisher Site  Google Scholar
 Y.C. Wang, Z.X. Yang, and N.Y. Deng, “Support vector machine prediction of enzyme function with conjoint triad feature and hierarchical context,” BMC Systems Biology, vol. 5, no. 1, article S6, 2011. View at: Publisher Site  Google Scholar
 O. L. Mangasarian and E. W. Wild, “Multisurface proximal support vector machine classification via generalized eigenvalues,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 28, no. 1, pp. 69–74, 2006. View at: Publisher Site  Google Scholar
 Jayadeva, R. Khemchandani, and S. Chandra, “Twin support vector machines for pattern classification,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 29, no. 5, pp. 905–910, 2007. View at: Publisher Site  Google Scholar
 Y.H. Shao, C.H. Zhang, X.B. Wang, and N.Y. Deng, “Improvements on twin support vector machines,” IEEE Transactions on Neural Networks, vol. 22, no. 6, pp. 962–968, 2011. View at: Publisher Site  Google Scholar
 M. Arun Kumar and M. Gopal, “Least squares twin support vector machines for pattern classification,” Expert Systems with Applications, vol. 36, no. 4, pp. 7535–7543, 2009. View at: Publisher Site  Google Scholar
 M. Arun Kumar, R. Khemchandani, M. Gopal, and S. Chandra, “Knowledge based least squares twin support vector machines,” Information Sciences, vol. 180, no. 23, pp. 4606–4618, 2010. View at: Publisher Site  Google Scholar
 Y.H. Shao, N.Y. Deng, and Z.M. Yang, “Least squares recursive projection twin support vector machine for classification,” Pattern Recognition, vol. 45, no. 6, pp. 2299–2307, 2012. View at: Publisher Site  Google Scholar  Zentralblatt MATH
 S. Gao, Q. Ye, and N. Ye, “1Norm least squares twin support vector machines,” Neurocomputing, vol. 74, no. 17, pp. 3590–3597, 2011. View at: Publisher Site  Google Scholar
 S. Ghorai, A. Mukherjee, and P. K. Dutta, “Nonparallel plane proximal classifier,” Signal Processing, vol. 89, no. 4, pp. 510–522, 2009. View at: Publisher Site  Google Scholar  Zentralblatt MATH
 M. A. Kumar and M. Gopal, “Application of smoothing technique on twin support vector machines,” Pattern Recognition Letters, vol. 29, no. 13, pp. 1842–1848, 2008. View at: Publisher Site  Google Scholar
 X. Peng, “A $v$twin support vector machine ($v$TSVM) classifier and its geometric algorithms,” Information Sciences, vol. 180, no. 20, pp. 3863–3875, 2010. View at: Publisher Site  Google Scholar  MathSciNet
 X. Peng, “TSVR: an efficient twin support vector machine for regression,” Neural Networks, vol. 23, no. 3, pp. 365–372, 2010. View at: Publisher Site  Google Scholar
 Z. Wang, J. Chen, and M. Qin, “Nonparallel planes support vector machine for multiclass classification,” in Proceedings of the International Conference on Logistics Systems and Intelligent Management (ICLSIM '10), vol. 1, pp. 581–585, January 2010. View at: Publisher Site  Google Scholar
 H. Cong, C. Yang, and X. Pu, “Efficient speaker recognition based on multiclass Twin Support Vector Machines and GMMs,” in Proceedings of the IEEE International Conference on Robotics, Automation and Mechatronics (RAM '08), pp. 348–352, September 2008. View at: Publisher Site  Google Scholar
 Z.X. Yang, Y.H. Shao, and X.S. Zhang, “Multiple birth support vector machine for multiclass classification,” Neural Computing and Applications, vol. 22, no. 1, pp. 153–161, 2013. View at: Publisher Site  Google Scholar
 A. N. Tikhonov and V. Y. Arsenin, Solutions of IllPosed Problems, John Wiley & Sons, New York, NY, USA, 1977.
 O. L. Mangasarian, Non Linear Programming, SIAM, Philadelphia, Pa, USA, 1994.
 O. L. Mangasarian and D. R. Musicant, “Successive overrelaxation for support vector machines,” IEEE Transactions on Neural Networks, vol. 10, no. 5, pp. 1032–1037, 1999. View at: Publisher Site  Google Scholar
 I. W.H. Tsang, A. Kocsor, and J. T.Y. Kwok, “Largescale maximum margin discriminant analysis using core vector machines,” IEEE Transactions on Neural Networks, vol. 19, no. 4, pp. 610–624, 2008. View at: Publisher Site  Google Scholar
 G. H. Golub and C. F. van Loan, Matrix Computations, The John Hopkins University Press, 1996. View at: MathSciNet
 Y. J. Lee and O. L. Mangasarian, Rsvm: Reduced Support Vector Machines, 2001.
 Y.J. Lee and S.Y. Huang, “Reduced support vector machines: a statistical theory,” IEEE Transactions on Neural Networks, vol. 18, no. 1, pp. 1–13, 2007. View at: Publisher Site  Google Scholar
 MATLAB, The MathWorks, 2007, http://www.mathworks.com.
 C. C. Chang and C. J. Lin, “LIBSVM: a library for support vector machines,” 2001, http://www.csie.ntu.edu.tw/~cjlin/. View at: Google Scholar
 R. O. Duda, P. E. Hart, and D. G. Stork, Pattern Classification, John Wiley & Sons, New York, NY, USA, 2nd edition, 2001. View at: MathSciNet
 C.W. Hsu and C.J. Lin, “A comparison of methods for multiclass support vector machines,” IEEE Transactions on Neural Networks, vol. 13, no. 2, pp. 415–425, 2002. View at: Publisher Site  Google Scholar
 C. L. Blake and C. J. Merz, “UCI Repository for Machine Learning Databases,” 1998, http://www.ics.uci.edu/~mlearn/MLRepository.html. View at: Google Scholar
Copyright
Copyright © 2015 ZhiXia Yang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.