1 / 27

Support Vector Machines

Support Vector Machines. H. Clara Pong Julie Horrocks 1 , Marianne Van den Heuvel 2 ,Francis Tekpetey 3, B. Anne Croy 4. 1 Mathematics & Statistics, University of Guelph, 2 Biomedical Sciences, University of Guelph, 3 Obstetrics and Gynecology, University of Western Ontario,

irisa
Télécharger la présentation

Support Vector Machines

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Support Vector Machines H. Clara Pong Julie Horrocks1, Marianne Van den Heuvel2,Francis Tekpetey3, B. Anne Croy4. 1 Mathematics & Statistics, University of Guelph, 2 Biomedical Sciences, University of Guelph, 3 Obstetrics and Gynecology, University of Western Ontario, 4 Anatomy & Cell Biology, Queen’s University

  2. Outline • Background • Separating Hyper-plane & Basis Expansion • Support Vector Machines • Simulations • Remarks

  3. CD56 bright cells Background • Motivation • The IVF (In-Vitro Fertilization) project • 18 infertile women • each undergoing the IVF treatment • Outcome (Outputs, Y’s) : Binary (pregnancy) • Predictor (Inputs, X’s): Longitudinal data (adhesion)

  4. Background • Classification methods • Relatively new method: Support Vector Machines • V. Vapnik: first proposed in 1979 • Maps input space into a high dimensional feature space • Constructs a linear classifier in the new feature space • Traditional method: Discriminant Analysis • R.A. Fisher: 1936 • Classify according to the values from the discriminant functions • Assumption: the predictors X in a given class has a Multi-Normal distribution.

  5. f(X) = β0 +βTX =0 A: f(X)>0 B: f(X)<0 Separating Hyper-plane Suppose there are 2 classes (A, B) • y = 1 for group A, y = -1 for group B. Let a hyper-plane be defined as f(X) = β0 +βTX = 0 then f(X) is the decision boundary that separates the two groups. f(X) = β0 +βTX > 0 for X Є A f(X) = β0 +βTX < 0 for X Є B Given X0 Є A, misclassified when f(X0 ) < 0. Given X0 Є B , misclassified when f(X0 ) > 0.

  6. f(X) = β0 +βTX =0 Separating Hyper-plane The perceptron learning algorithm search for a hyper-plane that minimizes the distance of misclassified points to the decision boundary. However this does not provide a unique solution.

  7. C C f(X) = β0* +β*TX = 0 Optimal Separating Hyper-plane Let C be the distance of the closest point from the two groups to the hyper-plane. The Optimal Separating hyper-plane is the unique separating hyper-plane f(X) = β0* +β*TX = 0, where (β*0 ,β*T) maximizes C.

  8. Dual LaGrange problem: f(X) = β0* +β*TX = 0 C C (the support vectors) Optimal Separating Hyper-plane Maximization Problem: Subjects to 1. αi [yi (xiTβ+ β0) -1] = 0 2. αi ≥ 0 all i=1…N 3. β = Σi=1..Nαi yixi 4. Σi=1..Nαi yi = 0 5. The Kuhn Tucker Conditions f(X) only depends on the xi’s where αi ≠ 0

  9. f(X) = β0* +β*TX = 0 C C (the support vectors) Optimal Separating Hyper-plane

  10. x1x2 x2 + x2 x1 x1 + Basis Expansion Suppose there are p inputs X=(x1 … xp) Let hk(X) be a transformation that maps X from RpR. hk(X) is called the basis function. H = {h1(X), … ,hm(X)} is the basis of a new feature space (dim=m) Example: X=(x1,x2) H = {h1(X), h2(X),h3(X)} h1(X) = h1(x1,x2) = x1, h2(X) = h2(x1,x2) =x2, h3(X) = h3(x1,x2) =x1x2 X_new = H(X)= (x1, x2, x1x2)

  11. Separable Case:all points are outside of the margins The classification rule is the sign of the decision function. C C f(X) = β0* +β*TX = 0 Support Vector Machines The optimal hyper-plane {X| f(X) = β0* +β*TX=0 }. f(X) = β0* +β*TX is called the Support Vector Classifier.

  12. yif(Xi) C Si Support Vector Machines Non-separable Case:training data is non-separable. Hyper-plane: {X| f(X) = β0 +βTX = 0 } f(X) = β0 +βTX = 0 Xi crosses the margin of its group when C – yi f(Xi) > 0. Si = C – yi f(Xi) when Xi crosses the margin and it’s zero when Xi outside. LetξiC =Si, ξi is the proportional of C that the prediction has crossed the margin. Misclassification occurs when Si > C (ξi > 1).

  13. Maximization Problem: Dual LaGrange problem: (non-separable case) s.t. . 0≤ αi ≤ ζ , Σ αi yi = 0 Subjects to 1. αi [yi (xiTβ+ β0) –(1-ξi)] = 0 2. vi ≥ 0 all i=1…N 3. β = Σ αiyixi 4. The Kuhn Tucker Conditions Support Vector Machines The overall misclassification is Σξi , and is bounded by δ.

  14. The linear classifier becomes Dual LaGrange problem: Support Vector Machines SVM search for an optimal hyper-plane in a new feature space where the data are more separate. Suppose H = {h1(X), … ,hm(X)} is the basis for the new feature space F. All elements in the new feature space is a linear basis expansion of X.

  15. Kernel: For example: This implies Support Vector Machines The kernel and the basis transformation define one another.

  16. The most common kernels: 1. dth Degree Polynomial: 2. Radial Basis: 3. Neural Network: Support Vector Machines Dual LaGrange function: This shows the basis transformation in SVM does not need to be define explicitly.

  17. Simulations • 3 cases • 100 simulations per case • Each simulation consists of 200 points • 100 points from each group • Input space: 2 dimensional • Output: 0 or 1 (2 groups) • Half of the points are randomly selected as the training set. X=(x1,x2), Y є {0,1}

  18. Black ~ group 0 Red ~ group 1 Simulations Case 1 (Normal with same covariance matrix)

  19. Simulations Case 1

  20. Black ~ group 0 Red ~ group 1 Simulations Case 2 (Normal with unequal covariance matrixes)

  21. Simulations Case 2

  22. Black ~ group 0 Red ~ group 1 Simulations Case 3 (Non-normal)

  23. Simulations Case 3

  24. Simulations Paired t-test for differences in misclassifications Ho: mean different = 0; Ha: mean different ≠ 0 Case 1 mean different (LDA - SVM) = - 0.41 , se = 0.3877 t = -1.057, p-value = 0.29 (insignificant) Case 2 mean different (QDA - SVM) = -1.96 , se = 0.4170 t = -4.70, p-value = 8.42e-06 (significant) Case 3 mean different (QDA - SVM) = 2, sd= 0.4218 t = 4.74, p-value = 7.13e-06 (significant)

  25. Remarks Support Vector Machines • Maps the original input space onto a feature space of higher dimension • No assumption on the distributions of X’s Performance • The performances of Discriminant Analysis and SVM are similar (when (X|Y) has a Normal distribution and share the same Σ) • Discriminant Analysis has a better performance (when the covariance matrices for the two groups are different) • SVM has a better performance (when the input (X) violated the distribution assumption)

  26. Reference • N. Cristianini, and J. Shawe-Taylor An introduction to Support Vector Machines and other kernel-based learning methods. New York: Cambridge University Press, 2000. • J. Friedman, T. Hastie, and R. Tibshirani The Elements of Statistical Learning. NewYork: Springer, 2001. • D. Meyer, C. Chang, and C. Lin.R Documentation: Support Vector Machines.http://www.maths.lth.se/help/R/.R/library/e1071/html/svm.html Last updated: March 2006 • H. Planatscher and J. Dietzsch. SVM-Tutorial using R (e1071-package) http://www.potschi.de/svmtut/svmtut.htm • M. Van Den Heuvel, J. Horrocks, S. Bashar, S. Taylor, S. Burke, K. Hatta, E. Lewis, and A. Croy. Menstrual Cycle Hormones Induce Changes in Functional Interac-tions Between Lymphocytes and Endothelial Cells. Journal of Clinical Endocrinology and Metabolism, 2005.

  27. Thank You !

More Related