110 likes | 226 Vues
This document provides a comprehensive overview of the dynamics of learning, emphasizing both single-agent learning theory and the emergence of distributed adaptation through collective learning. It discusses fundamental constraints in learning, measurement of cognitive capacity, and prediction in complex environments. Relevant mathematical foundations such as causal states and optimal predictors (ε-machines) are introduced. Additionally, it raises open questions regarding learning agents, interactions, and evolutionary dynamics. As part of the research agenda, it explores the implications for computational mechanics and agent-based models.
E N D
Dynamics of Learning & Distributed AdaptationPI: James P. Crutchfield, Santa Fe InstituteAgent-Based Computing Grantee Meeting3-5 October 2000 Dynamics of Learning: Single-agent learning theory Emergence of Distributed Adaptation: Agent-collective learning theory Related Projects: Computational Mechanics www.santafe.edu/projects/CompMech Evolutionary Dynamics www.santafe.edu/~evca Network Dynamics Program discuss.santafe.edu/dynamics Coyote: Multiprocessor for Agent-based Computing
Computational Mechanics:The Learning Channel TLC: Adaptation of Communication Channel What are fundamental constraints on learning? How to measure environmental structure? How to measure “cognitive” capacity of learning agents? How much data for a given complexity of inferred model?
Computational Mechanics:Preliminaries Observations: s = ss Past Future: … s-Ls-L+1…s-1s0|s1…s L-1sL … Probabilities: Pr(s), Pr(s), Pr(s) Uncertainty: Entropy H[P] = -i pi log pi [bits] Prediction error: Entropy Rate h = H[Pr(si|si-1si-2si-3…)] Information transmitted to future: Excess Entropy E = H[Pr(s)/ (Pr(s)Pr(s))] Measure of independence: Is Pr(s)=Pr(s)Pr(s)? Describes information in “raw” sequence blocks
Computational Mechanics:Mathematical Foundations Casual state = Condition of knowledge about future -Machines = {Causal states, Transitions} Optimality Theorem: -Machines are optimal predictors of environment. Minimality Theorem: Of the optimal predictors, -Machines are smallest. Uniqueness Theorem: Up to isomorphism, an -Machine is unique. The Point: Discovering an -Machine is the goal for any learning process. Practicalities may preclude this, but this is the goal. (w/ DP Feldman/CR Shalizi)
Computational Mechanics: Why Model? Structural Complexity of Information Source C = H[Pr(S)], S = {Casual states} Uses: Environ’l complexity: Amount/kind of relevant structure Agent’s inferential capacity: Sophistication of models? Theorem: E C Conclusion: Build models vs. storing only E bits of history. Raw sequence blocks do not allow optimal prediction, only E bits of mutual information in blocks. Optimal prediction requires larger model: 2C, not 2E. Explicit: 1D Range-R Ising spin system: C =E+Rh.
Dynamics of Learning: The Aha! Effect Learning complex environments (w/ C Douglas) Learning paradigm Three phases Memorization Aha! Refinement
Dynamics of Learning:Hierarchical Modeling Computation at the Onset of Chaos Onset of chaos leads to infinite -machine Learn the higher level representation Go from series of DFAs to Stack Automaton
Dynamics of Learning: Some Open Questions Learning agents Dynamical systems view of learning as a process whose behavior is predictive model building Define and measure agent “cognitive” abilities Development math’lly analyzable and simulatable models What state-space structures are responsible for learning? E.g., Basins = robust memories; bifurcations = adaptation; models = attractor-basin portrait in subspace; … Robot collectives Group versus individual function Define and measure degree of cooperation Agent collective functioning versus communication topologies
Evolutionary Dynamics Research Mathematical Analysis Epochal evolution Fitness barrier crossing: neutral paths v. fitness valleys? Optimal evolutionary search w/ E van Nimwegen (Dissn@SFI, Fall ‘99) Evolving Cellular Automata Population dynamics Embedded particle computation in CAs w/ W Hordijk (Dissn@SFI, Fall ‘99), M Mitchell, L Pagie, C Shalizi
Structure and DynamicsinComplex Interactive Networks Research Areas Network Structure Network Dynamics Hierarchical and Heterarchical Networks Components: Annual Workshop SFI-Intel Post-Doctoral Fellow Visitor Program Multiprocessor JPC-DW Individual Research Intel (BusNet) 3 years (JPC & DW)
Coyote: SFI’s Beowulf A Supercomputer for Complex Adaptive Systems • Cheap Off-the-Shelf Technology (“Piles of PCs”) • 64 Compute Nodes (128 CPUs), expandable • Fast Network Interconnect (Cisco Gigabit switch) • Physical: Gatehouse room (power/cooling retrofit) • General Availability: Summer Y2K • Team: • Lolly: Cluster Administration/Maintenance • JPC: Coordination, System Architecture • Tim: Node and Network Architecture • Alex: Parallel, Distributed Code Development