1 / 29

BLiNQ MEDIA Praneeth Vepakomma Senior Data Scientist

Generalization in Supervised Machine Learning. BLiNQ MEDIA Praneeth Vepakomma Senior Data Scientist. Hypothetical Knapsack of Coins:. Copper and Gold Coins Total number of coins is fixed and is a large sample. Capture-Recapture What is the proportion of Gold coins?.

beate
Télécharger la présentation

BLiNQ MEDIA Praneeth Vepakomma Senior Data Scientist

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Generalization in Supervised Machine Learning BLiNQ MEDIA Praneeth Vepakomma Senior Data Scientist

  2. Hypothetical Knapsack of Coins: • Copper and Gold Coins • Total number of coins is fixed and is a large sample. • Capture-Recapture • What is the proportion of Gold coins? • Copper and Gold Coins • Total number of coins is variable and is a large sample. • Capture-Recapture • What is the proportion of Gold coins?

  3. BASIC ML/STAT TERMINOLOGY:

  4. 190 Years after Gauss, the core problem of prediction remains an active problem : Then: Now:

  5. 190 Years after Gauss, the core problem of prediction remains an active problem : Find a mapping♯ from the features: is a list of parameters, required to represent the function #Approximation

  6. What is Supervised Learning? Loss Function Existing Features Known Labels Assumptions Loss Function Unavailable Features UnknownLabels

  7. Evaluating the Learned Function: • Loss Function quantifies the error in the approximation. • Learn a mapping by optimizing the loss. Example:

  8. Predictions with varying parameters:

  9. Predictions with varying parameters:

  10. How do we generalize?

  11. Generalization and Predictability Empirical Risk Minimization: • Empirical Risk is the average (expected) loss on seen data. • True Risk is the expected risk on the process generating the X,Y pairs. True Risk Minimization:

  12. PARAMETRIC CHARACTERIZATION OF THE MAPPING : • 2d-Linear function: Slope, Intercept • Cubic Spline: Number of knots, Location of Knots • Nearest-Neighbor regression: Number of neighbors • Lasso: L1-L2 Weights • Support Vector Machines: Kernel width, Margin Length • Random Forests: Resampling sample size

  13. Long list of available Supervised Learning Techniques. • Most of the techniques have tuning parameters. • We can minimize out-of-sample performance by tuning the technique with optimal parameters. • Tuning can be performed by cross-validation over a discrete grid of parameter combinations.

  14. CURSE OF DIMENSIONALITY- Flat World-10D World:

  15. CURSE OF DIMENSIONALITY- Flat World-10D World:

  16. CURSE OF DIMENSIONALITY- Flat World-10D World:

  17. CURSE OF DIMENSIONALITY- Let us validate:

  18. Structural Risk Minimization via Regularization:

  19. Geneology?

  20. Brief Description Technology Overview Hiring (What we’re looking for) http://blinqmedia.com/contact/job-openings/

  21. Lets work with Abalone

  22. Thank You!

More Related