1 / 62

Computational Mathematics for Large-scale Data Analysis

Explore the use of computational mathematics in analyzing large-scale datasets in the field of astrophysics. Discover how galaxies evolved, the nature of the early universe, the existence of dark energy, and the validity of our current model. Utilize statistical and machine learning techniques to handle the challenges posed by massive survey datasets.

susiea
Télécharger la présentation

Computational Mathematics for Large-scale Data Analysis

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Computational Mathematicsfor Large-scale Data Analysis Alexander Gray Georgia Institute of Technology College of Computing FASTlab: Fundamental Algorithmic and Statistical Tools Laboratory

  2. The FASTlabFundamental Algorithmic and Statistical Tools Laboratory • Arkadas Ozakin: Research scientist, PhD Theoretical Physics • Dong Ryeol Lee: PhD student, CS + Math • Ryan Riegel: PhD student, CS + Math • Parikshit Ram: PhD student, CS + Math • William March: PhD student, Math + CS • James Waters: PhD student, Physics + CS • Hua Ouyang: PhD student, CS • Sooraj Bhat: PhD student, CS • Ravi Sastry: PhD student, CS • Long Tran: PhD student, CS • Michael Holmes: PhD student, CS + Physics (co-supervised) • Nikolaos Vasiloglou: PhD student, EE (co-supervised) • Wei Guan: PhD student, CS (co-supervised) • Nishant Mehta: PhD student, CS (co-supervised) • Wee Chin Wong: PhD student, ChemE (co-supervised) • Abhimanyu Aditya: MS student, CS • Yatin Kanetkar: MS student, CS • Praveen Krishnaiah: MS student, CS • Devika Karnik: MS student, CS • Prasad Jakka: MS student, CS

  3. Instruments Data: CMB Maps Exponential growth in dataset sizes [Science, Szalay & J. Gray, 2001] 1990 COBE 1,000 2000 Boomerang 10,000 2002 CBI 50,000 2003 WMAP 1 Million 2008 Planck 10 Million Data: Local Redshift Surveys Data: Angular Surveys 1986 CfA 3,500 1996 LCRS 23,000 2003 2dF 250,000 2005 SDSS 800,000 1970 Lick 1M 1990 APM 2M 2005 SDSS 200M 2010 LSST 2B

  4. How did galaxies evolve? • What was the early universe like? • Does dark energy exist? • Is our model (GR+inflation) right? Astrophysicist R. Nichol, Inst. Cosmol. Gravitation A. Connolly, U. Pitt Physics C. Miller, NOAO R. Brunner, NCSA G. Djorgovsky, Caltech G. Kulkarni, Inst. Cosmol. Gravitation D. Wake, Inst. Cosmol. Gravitation R. Scranton,U. Pitt Physics M. Balogh, U. Waterloo Physics I. Szapudi, U. Hawaii Inst. Astronomy G. Richards, Princeton Physics A. Szalay, Johns Hopkins Physics Machine learning/ statistics guy

  5. How did galaxies evolve? • What was the early universe like? • Does dark energy exist? • Is our model (GR+inflation) right? Astrophysicist • Kernel density estimator • n-point spatial statistics • Nonparametric Bayes classifier • Support vector machine • Nearest-neighbor statistics • Gaussian process regression • Hierarchical clustering O(N2) O(Nn) R. Nichol, Inst. Cosmol. Grav. A. Connolly, U. Pitt Physics C. Miller, NOAO R. Brunner, NCSA G. Djorgovsky,Caltech G. Kulkarni, Inst. Cosmol. Grav. D. Wake, Inst. Cosmol. Grav. R. Scranton,U. Pitt Physics M. Balogh, U. Waterloo Physics I. Szapudi, U. Hawaii Inst. Astro. G. Richards, Princeton Physics A. Szalay, Johns Hopkins Physics O(N2) O(N2) O(N2) O(N3) O(cDT(N)) Machine learning/ statistics guy

  6. How did galaxies evolve? • What was the early universe like? • Does dark energy exist? • Is our model (GR+inflation) right? Astrophysicist • Kernel density estimator • n-point spatial statistics • Nonparametric Bayes classifier • Support vector machine • Nearest-neighbor statistics • Gaussian process regression • Hierarchical clustering O(N2) O(Nn) R. Nichol, Inst. Cosmol. Grav. A. Connolly, U. Pitt Physics C. Miller, NOAO R. Brunner, NCSA G. Djorgovsky, Caltech G. Kulkarni, Inst. Cosmol. Grav. D. Wake, Inst. Cosmol. Grav. R. Scranton,U. Pitt Physics M. Balogh, U. Waterloo Physics I. Szapudi, U. Hawaii Inst. Astro. G. Richards, Princeton Physics A. Szalay, Johns Hopkins Physics O(N2) O(N3) O(N2) O(N3) O(N3) Machine learning/ statistics guy

  7. How did galaxies evolve? • What was the early universe like? • Does dark energy exist? • Is our model (GR+inflation) right? Astrophysicist • Kernel density estimator • n-point spatial statistics • Nonparametric Bayes classifier • Support vector machine • Nearest-neighbor statistics • Gaussian process regression • Hierarchical clustering O(N2) O(Nn) R. Nichol, Inst. Cosmol. Grav. A. Connolly, U. Pitt Physics C. Miller, NOAO R. Brunner, NCSA G. Djorgovsky, Caltech G. Kulkarni, Inst. Cosmol. Grav. D. Wake, Inst. Cosmol. Grav. R. Scranton,U. Pitt Physics M. Balogh, U. Waterloo Physics I. Szapudi, U. Hawaii Inst. Astro. G. Richards, Princeton Physics A. Szalay, Johns Hopkins Physics O(N2) O(N3) O(N2) O(N3) O(N3) But I have 1 million points Machine learning/ statistics guy

  8. Uncertainty quantification Many types – all generally lead to more computational expense: • Uncertainty of statistical model: cross-validation, bootstrap • Uncertainty of simulation: run many times, estimate a model each time • Uncertainty of simulation vs. real data: expensive spatial statistics

  9. How to do Stats/Machine Learning on Very Large Survey Datasets? • Choose the appropriate statistical task and method for the scientific question • Use the fastest algorithm and data structure for the statistical method • Apply method to the data!

  10. How to do Stats/Machine Learning on Very Large Survey Datasets? • Choose the appropriate statistical task and method for the scientific question • Use the fastest algorithm and data structure for the statistical method • Apply method to the data!

  11. 10 data analysis problems, and scalable tools we’d like for them • Querying (e.g. characterizing a region of space): • spherical range-search O(N) • orthogonal range-search O(N) • k-nearest-neighbors O(N) • all-k-nearest-neighbors O(N2) • Density estimation (e.g. comparing galaxy types): • mixture of Gaussians • kernel density estimation O(N2) • L2 density tree [Ram and Gray in prep] • manifold kernel density estimation O(N3) [Ozakin and Gray 2008, to be submitted] • hyper-kernel density estimation O(N4) [Sastry and Gray 2008, submitted]

  12. 10 data analysis problems, and scalable tools we’d like for them 3. Regression (e.g. photometric redshifts): • linear regression O(D2) • kernel regression O(N2) • Gaussian process regression/kriging O(N3) 4. Classification (e.g. quasar detection, star-galaxy separation): • k-nearest-neighbor classifier O(N2) • nonparametric Bayes classifier O(N2) • support vector machine (SVM) O(N3) • non-negative SVM O(N3) [Guan and Gray, in prep] • false-positive-limiting SVM O(N3) [Sastry and Gray, in prep] • separation map O(N3) [Vasiloglou, Gray, and Anderson 2008, submitted]

  13. 10 data analysis problems, and scalable tools we’d like for them • Dimension reduction (e.g. galaxy or spectra characterization): • principal component analysis O(D2) • non-negative matrix factorization • kernel PCA O(N3) • maximum variance unfolding O(N3) • co-occurrence embedding O(N3) [Ozakin and Gray, in prep] • rank-based manifolds O(N3) [Ouyang and Gray 2008, ICML] • isometric non-negative matrix factorization O(N3) [Vasiloglou, Gray, and Anderson 2008, submitted] • Outlier detection (e.g. new object types, data cleaning): • by density estimation, by dimension reduction • by robust Lp estimation [Ram, Riegel and Gray, in prep]

  14. 10 data analysis problems, and scalable tools we’d like for them 7. Clustering (e.g. automatic Hubble sequence) • by dimension reduction, by density estimation • k-means • mean-shift segmentation O(N2) • hierarchical clustering (“friends-of-friends”) O(N3) 8. Time series analysis (e.g. asteroid tracking, variable objects): • Kalman filter O(D2) • hidden Markov model O(D2) • trajectory tracking O(Nn) • Markov matrix factorization [Tran, Wong, and Gray 2008, submitted] • functional independent component analysis [Mehta and Gray 2008, submitted]

  15. 10 data analysis problems, and scalable tools we’d like for them 9. Feature selection and causality (e.g. which features predict star/galaxy) • LASSO regression • L1 SVM • Gaussian graphical model inference and structure search • discrete graphical model inference and structure search • 0-1 feature-selecting SVM [Guan and Gray, in prep] • L1 Gaussian graphical model inference and structure search [Tran, Lee, Holmes, and Gray, in prep] 10. 2-sample testing and matching (e.g. cosmological validation, multiple surveys): • minimum spanning tree O(N3) • n-point correlation O(Nn) • bipartite matching/Gaussian graphical model inference O(N3) [Waters and Gray, in prep]

  16. How to do Stats/Machine Learning on Very Large Survey Datasets? • Choose the appropriate statistical task and method for the scientific question • Use the fastest algorithm and data structure for the statistical method • Apply method to the data!

  17. Core computational problems What are the basic mathematical operations, or bottleneck subroutines, can we focus on developing fast algorithms for?

  18. Core computational problemsAggregations,GNPs,graphs,linear algebra,optimization, integration • Querying:nearest-neighbor, sph range-search, ortho range-search, all-nn • Density estimation:kernel density estimation, mixture of Gaussians • Regression:linear regression,kernel regression,Gaussian process regression • Classification:nearest-neighbor classifier, nonparametric Bayes classifier,support vectormachine • Dimension reduction:principal component analysis,non-negative matrix factorization, kernelPCA, maximum variance unfolding • Outlier detection: by robust L2 estimation, by density estimation, by dimension reduction • Clustering:k-means, mean-shift, hierarchical clustering (“friends-of-friends”), by dimension reduction, by density estimation • Time series analysis:Kalman filter,hiddenMarkov model, trajectory tracking • Feature selection and causality: LASSOregression, L1 support vectormachine, Gaussian graphical models, discrete graphical models • 2-sample testing and matching:n-point correlation, bipartite matching

  19. Simple example: spherical range search. How can we compute this efficiently? kd-trees: most widely-used space-partitioning tree [Bentley 1975], [Friedman, Bentley & Finkel 1977],[Moore & Lee 1995]

  20. A kd-tree: level 1

  21. A kd-tree: level 2

  22. A kd-tree: level 3

  23. A kd-tree: level 4

  24. A kd-tree: level 5

  25. A kd-tree: level 6

  26. Range-count recursive algorithm

  27. Range-count recursive algorithm

  28. Range-count recursive algorithm

  29. Range-count recursive algorithm

  30. Range-count recursive algorithm Pruned! (inclusion)

  31. Range-count recursive algorithm

  32. Range-count recursive algorithm

  33. Range-count recursive algorithm

  34. Range-count recursive algorithm

  35. Range-count recursive algorithm

  36. Range-count recursive algorithm

  37. Range-count recursive algorithm

  38. Range-count recursive algorithm Pruned! (exclusion)

  39. Range-count recursive algorithm

  40. Range-count recursive algorithm

  41. Range-count recursive algorithm fastest practical algorithm [Bentley 1975] our algorithms can use any tree

  42. How to do Stats/Machine Learning on Very Large Survey Datasets? • Choose the appropriate statistical task and method for the scientific question • Use the fastest algorithm and data structure for the statistical method • Apply method to the data!

  43. All-k-nearest neighbors • Applications: querying, adaptive density estimation first step [Budavari et al., in prep] • Naïve cost: O(N2) • Fastest algorithm: Generalized Fast Multipole Method, aka multi-tree methods • Technique: Use two trees simultaneously • [Gray and Moore 2001, NIPS; Riegel, Boyer and Gray, to be submitted] • O(N), exact

  44. Kernel density estimation • Applications: accurate density estimates, e.g. galaxy types [Balogh et al. 2004] • Naïve cost: O(N2) • Fastest algorithm: dual-tree fast Gauss transform, Monte Carlo multipole method • Technique: Hermite operators, Monte Carlo • [Lee, Gray and Moore 2005, NIPS; Lee and Gray 2006, UAI; Holmes, Gray and Isbell 2007, NIPS; Lee and Gray 2008, NIPS] • O(N), relative error guarantee

  45. Nonparametric Bayes classification (or KDA) • Applications: accurate classification, e.g. quasar detection [Richards et al. 2004,2009], [Scranton et al. 2005] • Naïve cost: O(N2) • Fastest algorithm: dual-tree bounding with hybrid tree expansion • Technique: bound each posterior • [Liu, Moore, and Gray 2004; Gray and Riegel 2004, CompStat; Riegel and Gray 2007, SDM] • O(N), exact

  46. Friends-of-friends • Applications: cluster-finding, 2-sample • Naïve cost: O(N3) • Fastest algorithm: dual-tree Boruvka algorithm • Technique: similar to all-k-nn • [March and Gray, to be submitted] • O(NlogN), exact

  47. n-point correlations • Applications: 2-sample, e.g. AGN [Wake et al. 2004], ISW [Giannantonio et al. 2006], 3pt validation [Kulkarni et al. 2007] • Naïve cost: O(Nn) • Fastest algorithm: n-tree algorithms • Technique: use n trees simultaneously • [Gray and Moore 2001, NIPS; Moore et al. 2001, Mining the Sky; Waters, Riegel and Gray, to be submitted] • O(Nlogn), exact

  48. 3-point runtime (biggest previous: 20K) VIRGO simulation data, N = 75,000,000 naïve: 5x109 sec. (~150 years) multi-tree: 55 sec. (exact) n=2: O(N) n=3: O(Nlog3) n=4: O(N2)

  49. Now fast!very fast as fast as possible (conjecture) • Querying:nearest-neighbor,sph range-search, ortho range-search, all-nn • Density estimation:kernel density estimation, mixture of Gaussians • Regression:linear regression, kernel regression,Gaussian process regression • Classification:nearest-neighbor classifier, nonparametric Bayes classifier, support vector machine • Dimension reduction:principal component analysis, non-negative matrix factorization,kernel PCA,maximum variance unfolding • Outlier detection: by robust L2 estimation • Clustering:k-means, mean-shift, hierarchical clustering (“friends-of-friends”) • Time series analysis:Kalman filter, hidden Markov model,trajectory tracking • Feature selection and causality: LASSO regression, L1 support vector machine, Gaussian graphical models, discrete graphical models • 2-sample testing and matching:n-point correlation, bipartite matching

  50. The end • You can now use many of the state-of-the-art data analysis methods on large datasets • MLPACK software • CAS Analytics in SQL Server • Computational astronomy workshop and large-scale data analysis workshop coming soon Alexander Grayagray@cc.gatech.edu (email is best; webpage sorely out of date)

More Related