1 / 25

Multivariate Analysis, TMVA, and Artificial Neural Networks

http://tmva.sourceforge.net/. Multivariate Analysis, TMVA, and Artificial Neural Networks. Matt Jachowski jachowski@stanford.edu. Multivariate Analysis. Techniques dedicated to analysis of data with multiple variables

manuelaa
Télécharger la présentation

Multivariate Analysis, TMVA, and Artificial Neural Networks

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. http://tmva.sourceforge.net/ Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu Matt Jachowski

  2. Multivariate Analysis • Techniques dedicated to analysis of data with multiple variables • Active field – many recently developed techniques rely on computational ability of modern computers Matt Jachowski

  3. Multivariate Analysis and HEP • Goal is to classify events as signal or background • Single event defined by several variables (energy, transverse momentum, etc.) • Use all the variables to classify the event • Multivariate analysis! Matt Jachowski

  4. Multivariate Analysis and HEP • Rectangular cuts optimization common Matt Jachowski

  5. Multivariate Analysis and HEP • Likelihood Estimator analysis also common • Use of more complicated methods (Neural Networks, Boosted Decision Trees) not so common (though growing) – why? • Difficult to implement • Physicists are skeptical of new methods Matt Jachowski

  6. Toolkit for Multivariate Analysis (TMVA) • ROOT-integrated software package with several MVA techniques • Automatic training, testing, and evaluation of MVA methods • Guidelines and documentation to describe methods for users – this isn’t a black box! Matt Jachowski

  7. Toolkit for Multivariate Analysis (TMVA) • Easy to configure methods • Easy to “plug-in” HEP data • Easy to compare different MVA methods Matt Jachowski

  8. TMVA in Action Matt Jachowski

  9. TMVA and Me • TMVA started in October 2005 • Still young • Very active group of developers • My involvement • Decorrelation for Cuts Method (mini project) • New Artificial Neural Network implementation (main project) Matt Jachowski

  10. Decorrelated Cuts Method • Some MVA methods suffer if data has linear correlations • i.e. Likelihood Estimator, Cuts • Linear correlations can be easily transformed away • I implemented this for the Cuts Method Matt Jachowski

  11. Decorrelated Cuts Method • Find the square root of the covariance matrix (C=C’C’) • Decorrelate the data • Apply cuts to decorrelated data Matt Jachowski

  12. Artificial Neural Networks (ANNs) • Robust non-linear MVA technique Matt Jachowski

  13. Matt Jachowski

  14. Training an ANN • Challenge is training the network • Like human brain, network learns from seeing data over and over again • Technical details:Ask me if you’re really interested Matt Jachowski

  15. MLP • MLP (Multi-Layer Perceptron) – my ANN implementation for TMVA • MLP is TMVA’s main ANN • MLP serves as base for any future ANN developments in TMVA Matt Jachowski

  16. MLP – Information & Statistics • Implemented in C++ • Object-Oriented • 4,000+ lines of code • 16 classes Matt Jachowski

  17. Acknowledgements • Joerg Stelzer • Andreas Hoecker • CERN • University of Michigan • Ford • NSF Matt Jachowski

  18. Questions?(I have lots of technical slides in reserve that I would be glad to talk about) Matt Jachowski

  19. Matt Jachowski

  20. Synapses and Neurons v0 y0 w0j v1 y1 vj yj w1j . . . wnj vn yn Matt Jachowski

  21. yj vj Synapses and Neurons Matt Jachowski

  22. Universal Approximation Theorem Every continuous function that maps intervals of real numbers to some output interval of real numbers can be approximated arbitrarily closely by a multi-layer perceptron with just one hidden layer (with non-linear activation functions). output inputs weights between hidden and output layer weights between input and hidden layer non-linear activation function bias Matt Jachowski

  23. x0 x1 y x2 x3 Training an MLP Training Event: Network: Matt Jachowski

  24. Training an MLP Adjust weights to minimize error (or an estimator that is some function of the error) Matt Jachowski

  25. Back-Propagation Algorithm Make correction in direction of steepest descent Corrections made to output layer first, propagated backwards Matt Jachowski

More Related