110 likes | 219 Views
Dynamics of Learning & Distributed Adaptation PI: James P. Crutchfield, Santa Fe Institute Agent-Based Computing Grantee Meeting 3-5 October 2000. Dynamics of Learning: Single-agent learning theory Emergence of Distributed Adaptation:
E N D
Dynamics of Learning & Distributed AdaptationPI: James P. Crutchfield, Santa Fe InstituteAgent-Based Computing Grantee Meeting3-5 October 2000 Dynamics of Learning: Single-agent learning theory Emergence of Distributed Adaptation: Agent-collective learning theory Related Projects: Computational Mechanics www.santafe.edu/projects/CompMech Evolutionary Dynamics www.santafe.edu/~evca Network Dynamics Program discuss.santafe.edu/dynamics Coyote: Multiprocessor for Agent-based Computing
Computational Mechanics:The Learning Channel TLC: Adaptation of Communication Channel What are fundamental constraints on learning? How to measure environmental structure? How to measure “cognitive” capacity of learning agents? How much data for a given complexity of inferred model?
Computational Mechanics:Preliminaries Observations: s = ss Past Future: … s-Ls-L+1…s-1s0|s1…s L-1sL … Probabilities: Pr(s), Pr(s), Pr(s) Uncertainty: Entropy H[P] = -i pi log pi [bits] Prediction error: Entropy Rate h = H[Pr(si|si-1si-2si-3…)] Information transmitted to future: Excess Entropy E = H[Pr(s)/ (Pr(s)Pr(s))] Measure of independence: Is Pr(s)=Pr(s)Pr(s)? Describes information in “raw” sequence blocks
Computational Mechanics:Mathematical Foundations Casual state = Condition of knowledge about future -Machines = {Causal states, Transitions} Optimality Theorem: -Machines are optimal predictors of environment. Minimality Theorem: Of the optimal predictors, -Machines are smallest. Uniqueness Theorem: Up to isomorphism, an -Machine is unique. The Point: Discovering an -Machine is the goal for any learning process. Practicalities may preclude this, but this is the goal. (w/ DP Feldman/CR Shalizi)
Computational Mechanics: Why Model? Structural Complexity of Information Source C = H[Pr(S)], S = {Casual states} Uses: Environ’l complexity: Amount/kind of relevant structure Agent’s inferential capacity: Sophistication of models? Theorem: E C Conclusion: Build models vs. storing only E bits of history. Raw sequence blocks do not allow optimal prediction, only E bits of mutual information in blocks. Optimal prediction requires larger model: 2C, not 2E. Explicit: 1D Range-R Ising spin system: C =E+Rh.
Dynamics of Learning: The Aha! Effect Learning complex environments (w/ C Douglas) Learning paradigm Three phases Memorization Aha! Refinement
Dynamics of Learning:Hierarchical Modeling Computation at the Onset of Chaos Onset of chaos leads to infinite -machine Learn the higher level representation Go from series of DFAs to Stack Automaton
Dynamics of Learning: Some Open Questions Learning agents Dynamical systems view of learning as a process whose behavior is predictive model building Define and measure agent “cognitive” abilities Development math’lly analyzable and simulatable models What state-space structures are responsible for learning? E.g., Basins = robust memories; bifurcations = adaptation; models = attractor-basin portrait in subspace; … Robot collectives Group versus individual function Define and measure degree of cooperation Agent collective functioning versus communication topologies
Evolutionary Dynamics Research Mathematical Analysis Epochal evolution Fitness barrier crossing: neutral paths v. fitness valleys? Optimal evolutionary search w/ E van Nimwegen (Dissn@SFI, Fall ‘99) Evolving Cellular Automata Population dynamics Embedded particle computation in CAs w/ W Hordijk (Dissn@SFI, Fall ‘99), M Mitchell, L Pagie, C Shalizi
Structure and DynamicsinComplex Interactive Networks Research Areas Network Structure Network Dynamics Hierarchical and Heterarchical Networks Components: Annual Workshop SFI-Intel Post-Doctoral Fellow Visitor Program Multiprocessor JPC-DW Individual Research Intel (BusNet) 3 years (JPC & DW)
Coyote: SFI’s Beowulf A Supercomputer for Complex Adaptive Systems • Cheap Off-the-Shelf Technology (“Piles of PCs”) • 64 Compute Nodes (128 CPUs), expandable • Fast Network Interconnect (Cisco Gigabit switch) • Physical: Gatehouse room (power/cooling retrofit) • General Availability: Summer Y2K • Team: • Lolly: Cluster Administration/Maintenance • JPC: Coordination, System Architecture • Tim: Node and Network Architecture • Alex: Parallel, Distributed Code Development