1 / 12

Feature Selection

Feature Selection.

tadita
Télécharger la présentation

Feature Selection

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Feature Selection • Not all features/attributes are relevant to the problem. Even some of them interfere and reduces accuracy. Such noisy features should be deleted in order to increase accuracy. In addition, dealing with larger number of features is costly, and therefore reducing dimensionality is rather important. • There are many methods used to select optimum number of features (e.g., backward or forward stepwise methods) • In genetics, there are thousands of genes measured for disease or condition. However, not all of them are associated with the disease or condition. It is important to identity a set of genes that are the most associated with the disease/condition, and then such small number of genes can be used as a drug target. Such approach is widely used in cancer area where associative genes are being identified.

  2. Feature Selection Optimal solution is to try all possible combinations of the features (=2m-1)

  3. Feature Selection Methods IEEE PAMI, 2000, 22(1):4-37

  4. Some other methods for feature selection available in MATLAB

  5. Feature Extraction and Projection Methods IEEE PAMI, 2000, 22(1):4-37

  6. Principalcomponentsanalysis In data sets with many variables, groups of variables often move together. One reason for this is that more than one variable might be measuring the same driving principle governing the behavior of the system. In many systems there are only a few such driving forces. But an abundance of instrumentation enables you to measure dozens of system variables. When this happens, you can take advantage of this redundancy of information. You can simplify the problem by replacing a group of variables with a single new variable. Principal components analysis is a quantitatively rigorous method for achieving this simplification. The method generates a new set of variables, called principal components. Each principal component is a linear combination of the original variables. All the principal components are orthogonal to each other, so there is no redundant information. The principal components as a whole form an orthogonal basis for the space of the data.

  7. There are an infinite number of ways to construct an orthogonal basis for several columns of data. • The first principal component is a single axis in space. When you project each observation on that axis, the resulting values form a new variable. And the variance of this variable is the maximum among all possible choices of the first axis. • The second principal component is another axis in space, perpendicular to the first. Projecting the observations on this axis generates another new variable. The variance of this variable is the maximum among all possible choices of this second axis. • The full set of principal components is as large as the original set of variables. But it is commonplace for the sum of the variances of the first few principal components to exceed 80% of the total variance of the original data. By examining plots of these few new variables, researchers often develop a deeper understanding of the driving forces that generated the original data.

More Related