1957–58 Beşiktaş J.K. season

From formulasearchengine
Revision as of 08:02, 22 August 2013 by en>Yobot (External links: WP:CHECKWIKI error fixes / special characters in sortkey fixed using AWB (9440))
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

My name is Jestine (34 years old) and my hobbies are Origami and Microscopy.

Here is my web site; http://Www.hostgator1centcoupon.info/ (support.file1.com) Least squares support vector machines (LS-SVM) are least squares versions of support vector machines (SVM), which are a set of related supervised learning methods that analyze data and recognize patterns, and which are used for classification and regression analysis. In this version one finds the solution by solving a set of linear equations instead of a convex quadratic programming (QP) problem for classical SVMs. Least squares SVM classifiers, were proposed by Suykens and Vandewalle.[1] LS-SVMs are a class of kernel-based learning methods.

From support vector machine to least squares support vector machine

Given a training set {xi,yi}i=1N with input data xin and corresponding binary class labels yi{1,+1}, the SVM[2] classifier, according to Vapnik’s original formulation, satisfies the following conditions:

The spiral data yi=1 for blue data point yi=1 for red data point
{wTϕ(xi)+b1,if yi=+1,wTϕ(xi)+b1,if yi=1.

Which is equivalent to

yi[wTϕ(xi)+b]1,i=1,,N,

where ϕ(x) is the nonlinear map from original space to the high (and possibly infinite) dimensional space.

Inseparable data

In case of such separating hyperplane does not exist, we introduce a so-called slack variable ξi such that

{yi[wTϕ(xi)+b]1ξi,i=1,,N,ξi0,i=1,,N.

According to the structural risk minimization principle, the risk bound is minimized by the following minimization problem:

minJ1(w,ξ)=12wTw+ci=1Nξi,
Subject to {yi[wTϕ(xi)+b]1ξi,i=1,,N,ξi0,i=1,,N,
The result of the SVM classifier

To solve this problem, we could construct the Lagrangian function:

L1(w,b,ξ,α,β)=12wTw+ci=1Nξi+i=1Nαi{yi[wTϕ(xi)+b]1+ξi}+i=1Nβiξi,

where αi0,βi0(i=1,,N) are the Lagrangian multipliers. The optimal point will in the saddle point of the Lagrangian function, and then we obtain

{L1w=0w=i=1Nαiyiϕ(xi),L1b=0i=1Nαiyi=0,L1ξi=00αic,i=1,,N.

By substituting w by its expression in the Lagrangian formed from the appropriate objective and constraints, we will get the following quadratic programming problem:

maxQ1(α)=12i,j=1NαiαjyiyjK(xi,xj)+i=1Nαi

where K(xi,xj)=ϕ(xi),ϕ(xj) is called the kernel function. Solving this QP problem subject to constraints in (8), we will get the hyperplane in the high dimensional space and hence the classifier in the original space.

Least squares SVM formulation

The least squares version of the SVM classifier is obtained by reformulating the minimization problem as:

minJ2(w,b,e)=μ2wTw+ζ2i=1Nec,i2,

subject to the equality constraints:

yi[wTϕ(xi)+b]=1ec,i,i=1,,N.

The least squares SVM (LS-SVM) classifier formulation above implicitly corresponds to a regression interpretation with binary targets yi=±1.

Using yi2=1, we have

i=1Nec,i2=i=1N(yiec,i)2=i=1Nei2=i=1N(yi(wTϕ(xi)+b))2,

with ei=yi(wTϕ(xi)+b). Notice, that this error would also make sense for least squares data fitting, so that the same end results holds for the regression case.

Hence the LS-SVM classifier formulation is equivalent to

J2(w,b,e)=μEW+ζED

with EW=12wTw and ED=12i=1Nei2=12i=1N(yi(wTϕ(xi)+b))2.

The result of the LS-SVM classifier

Both μ and ζ should be considered as hyperparamters to tune the amount of regularization versus the sum squared error. The solution does only depend on the ratio γ=ζ/μ, therefore the original formulation uses only γ as tuning parameter. We use both μ and ζ as parameters in order to provide a Bayesian interpretation to LS-SVM.

The solution of LS-SVM regressor will be obtained after we construct the Lagrangian function:

{L2(w,b,e,α)=J2(w,e)i=1Nαi{[wTϕ(xi)+b]+eiyi},=12wTw+γ2i=1Nei2i=1Nαi{[wTϕ(xi)+b]+eiyi},

where αi are the Lagrange multipliers. The conditions for optimality are

{L2w=0w=i=1Nαiϕ(xi),L2b=0i=1Nαi=0,L2ei=0αi=γei,i=1,,N,L2αi=0yi=wTϕ(xi)+b+ei,i=1,,N.

Elimination of w and e will yield a linear system instead of a quadratic programming problem:

[01NT1NΩ+γ1IN][bα]=[0Y],

with Y=[y1,,yN]T, 1N=[1,,1]T and α=[α1,,αN]T. Here, IN is an N×N identity matrix, and ΩN×N is the kernel matrix defined by Ωij=ϕ(xi)Tϕ(xj)=K(xi,xj).

Kernel function K

For the kernel function K(•, •) one typically has the following choices:

where d, c, σ, k and θ are constants. Notice that the Mercer condition holds for all c,σ+ and dN values in the polynomial and RBF case, but not for all possible choices of k and θ in the MLP case. The scale parameters c, σ and k determine the scaling of the inputs in the polynomial, RBF and MLP kernel function. This scaling is related to the bandwidth of the kernel in statistics, where it is shown that the bandwidth is an important parameter of the generalization behavior of a kernel method.

Bayesian interpretation for LS-SVM

A Bayesian interpretation of the SVM has been proposed by Smola et al. They showed that the use of different kernels in SVM can be regarded as defining different prior probability distributions on the functional space, as P[f]exp(βP^f2) . Here β>0 is a constant and P^ is the regularization operator corresponding to the selected kernel.

A general Bayesian evidence framework was developed by MacKay,[3][4][5] and MacKay has used it to the problem of regression, forward neural network and classification network. Provided data set D, a model 𝕄 with parameter vector w and a so-called hyperparameter or regularization parameter λ, Bayesian inference is constructed with 3 levels of inference:

  • In level 1, for a given value of λ, the first level of inference infers the posterior distribution of by Bayesian rule
p(w|D,λ,𝕄)p(D|w,𝕄)p(w|λ,𝕄)
  • The second level of inference determines the value of λ, by maximizing
p(λ|D,𝕄)p(D|λ,𝕄)p(λ|𝕄)
  • The third level of inference in the evidence framework ranks different models by examining their posterior probabilities
p(𝕄|D)p(D|𝕄)p(𝕄).

We can see that Bayesian evidence framework is a unified theory for learning the model and model selection. Kwok used the Bayesian evidence framework to interpret the formulation of SVM and model selection. And he also applied Bayesian evidence framework to support vector regression.

Now, given the data points {xi,yi}i=1N and the hyperparameters μ and ζ of the model 𝕄, the model parameters w and b are estimated by maximizing the posterior p(w,b|D,logμ,logζ,𝕄). Applying Bayes’ rule, we obtain:

p(w,b|D,logμ,logζ,𝕄)=p(D|w,b,logμ,logζ,𝕄)p(w,b|logμ,logζ,𝕄)p(D|logμ,logζ,𝕄).

Where p(D|logμ,logζ,𝕄) is a normalizing constant such the integral over all possible w and b is equal to 1. We assume w and b are independent of the hyperparameter ζ, and are conditional independent, i.e., we assume

p(w,b|logμ,logζ,𝕄)=p(w|logμ,𝕄)p(b|logσb,𝕄).

When σb, the distribution of b will approximate a uniform distribution. Furthermore, we assume w and b are Gaussian distribution, so we obtain the a priori distribution of w and b with σb to be:

p(w,b|logμ,)=(μ2π)nf2exp(μ2wTw)12πσbexp(b22σb)(μ2π)nf2exp(μ2wTw).

Here nf is the dimensionality of the feature space, same as the dimensionality of w.

The probability of p(D|w,b,logμ,logζ,𝕄) is assumed to depend only on w,b,ζ and 𝕄. We assume that the data points are independently identically distributed (i.i.d.), so that:

p(D|w,b,logζ,𝕄)=i=1Np(xi,yi|w,b,logζ,𝕄).

In order to obtain the least square cost function, it is assumed that the probability of a data point is proportional to:

p(xi,yi|w,b,logζ,𝕄)p(ei|w,b,logζ,𝕄).

A Gaussian distribution is taken for the errors ei=yi(wTϕ(xi)+b) as:

p(ei|w,b,logζ,𝕄)=ζ2πexp(ζei22).

It is assumed that the w and b are determined in such a way that the class centers m^ and m^+ are mapped onto the target -1 and +1, respectively. The projections wTϕ(x)+b of the class elements ϕ(x) follow a multivariate Gaussian distribution, which have variance 1/ζ.

Combining the preceding expressions, and neglecting all constants, Bayes’ rule becomes

p(w,b|D,logμ,logζ,𝕄)exp(μ2wTwζ2i=1Nei2)=exp(J2(w,b)).

The maximum posterior density estimates wMP and bMP are then be obtained by minimizing the negative logarithm of (26), so we arrive (10).

References

  1. Suykens, J.A.K.; Vandewalle, J. (1999) "Least squares support vector machine classifiers", Neural Processing Letters, 9 (3), 293-300.
  2. Vapnik, V. The nature of statistical learning theory. Springer-Verlag, New York, 1995
  3. MacKay, D.J.C. Bayesian Interpolation. Neural Computation, 4(3): 415-447, May 1992.
  4. MacKay, D.J.C. A practical Bayesian framework for backpropagation networks. Neural Computation, 4(3): 448-472, May 1992.
  5. MacKay, D.J.C. The evidence framework applied to classification networks. Neural Computation, 4(5): 720-736, Sept. 1992.

Bibliography

  • J. A. K. Suykens, T. Van Gestel, J. De Brabanter, B. De Moor, J. Vandewalle, Least Squares Support Vector Machines, World Scientific Pub. Co., Singapore, 2002. ISBN 981-238-151-1
  • Suykens J.A.K., Vandewalle J., Least squares support vector machine classifiers, Neural Processing Letters, vol. 9, no. 3, Jun. 1999, pp. 293–300.
  • Vladimir Vapnik. The Nature of Statistical Learning Theory. Springer-Verlag, 1995. ISBN 0-387-98780-0
  • MacKay, D. J. C., Probable networks and plausible predictions—A review of practical Bayesian methods for supervised neural networks. Network: Computation in Neural Systems, vol. 6, 1995, pp. 469–505.

External links

  • www.esat.kuleuven.be/sista/lssvmlab/ "Least squares support vector machine Lab (LS-SVMlab) toolbox contains Matlab/C implementations for a number of LS-SVM algorithms."
  • www.kernel-machines.org "Support Vector Machines and Kernel based methods (Smola & Schölkopf)."
  • www.gaussianprocess.org "Gaussian Processes: Data modeling using Gaussian Process priors over functions for regression and classification (MacKay, Williams)"
  • www.support-vector.net "Support Vector Machines and kernel based methods (Cristianini)"
  • dlib: Contains a least-squares SVM implementation for large-scale datasets.