1 / 48

20 Minute Quiz

20 Minute Quiz. For each of the questions, you can use text, diagrams, bullet points, etc. Give three processes that keep neural computation from proceeding faster than at millisecond scale. How does pre-natal activity dependent visual tuning work?

ayita
Télécharger la présentation

20 Minute Quiz

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. 20 Minute Quiz For each of the questions, you can use text, diagrams, bullet points, etc. • Give three processes that keep neural computation from proceeding faster than at millisecond scale. • How does pre-natal activity dependent visual tuning work? • How does the knee-jerk reflex work? Describe another similar reflex found in humans. • Identify two philosophical issues discussed in the assigned chapters of the M2M book.

  2. How does activity lead to structural change? • The brain (pre-natal, post-natal, and adult) exhibits a surprising degree of activity dependent tuning and plasticity. • To understand the nature and limits of the tuning and plasticity mechanisms we study • How activity is converted to structural changes (say the ocular dominance column formation) • It is centrally important • to arrive at biological accounts of perceptual, motor, cognitive and language learning • Biological Learning is concerned with this topic.

  3. Learning and Memory Declarative Fact based/Explicit Skill based Implicit Episodic Semantic Procedural Learning and Memory: Introduction facts about a situation general facts skills

  4. Skill and Fact Learning may involve different mechanisms • Certain brain injuries involving the hippocampal region of the brain render their victims incapable of learning any new facts or new situations or faces. • But these people can still learn new skills, including relatively abstract skills like solving puzzles. • Fact learning can be single-instance based. Skill learning requires repeated exposure to stimuli • subcortical structures like the cerebellum and basal ganglia seem to play a role in skill learning • Implications for Language Learning?

  5. Models of Learning • Hebbian ~ coincidence • Recruitment ~ one trial • Supervised ~ correction (backprop) • Reinforcement ~ delayed reward • Unsupervised ~ similarity

  6. Hebb’s Rule • The key idea underlying theories of neural learning go back to the Canadian psychologist Donald Hebb and is called Hebb’s rule. • From an information processing perspective, the goal of the system is to increase the strength of the neural connections that are effective.

  7. Hebb (1949) “When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that A’s efficiency, as one of the cells firing B, is increased” From: The organization of behavior.

  8. Hebb’s rule • Each time that a particular synaptic connection is active, see if the receiving cell also becomes active. If so, the connection contributed to the success (firing) of the receiving cell and should be strengthened. If the receiving cell was not active in this time period, our synapse did not contribute to the success the trend and should be weakened.

  9. strengthen weaken LTP and Hebb’s Rule • Hebb’s Rule: neurons that fire together wire together • Long Term Potentiation (LTP) is the biological basis of Hebb’s Rule • Calcium channels are the key mechanism

  10. Chemical realization of Hebb’s rule • It turns out that there are elegant chemical processes that realize Hebbian learning at two distinct time scales • Early Long Term Potentiation (LTP) • Late LTP • These provide the temporal and structural bridge from short term electrical activity, through intermediate memory, to long term structural changes.

  11. Calcium Channels Facilitate Learning • In addition to the synaptic channels responsible for neural signaling, there are also Calcium-based channels that facilitate learning. • As Hebb suggested, when a receiving neuron fires, chemical changes take place at each synapse that was active shortly before the event.

  12. Long Term Potentiation (LTP) • These changes make each of the winning synapses more potent for an intermediate period, lasting from hours to days (LTP). • In addition, repetition of a pattern of successful firing triggers additional chemical changes that lead, in time, to an increase in the number of receptor channels associated with successful synapses - the requisite structural change for long term memory. • There are also related processes for weakening synapses and also for strengthening pairs of synapses that are active at about the same time.

  13. Amygdala Hippocampus Temporal lobe LTP is found in the hippocampus Essential for declarative memory (Episodic Memory) In the temporal lobe Cylindrical Structure

  14. The Hebb rule is found with long term potentiation (LTP) in the hippocampus Schafer collateral pathway Pyramidal cells 1 sec. stimuli At 100 hz

  15. During normal low-frequency trans-mission, glutamate interacts with NMDA and non-NMDA (AMPA) and metabotropic receptors. With high-frequency stimulation

  16. Early and late LTP • (Kandel, ER, JH Schwartz and TM Jessell (2000) Principles of Neural Science. New York: McGraw-Hill.) • Experimental setup for demonstrating LTP in the hippocampus. The Schaffer collateral pathway is stimulated to cause a response in pyramidal cells of CA1. • Comparison of EPSP size in early and late LTP with the early phase evoked by a single train and the late phase by 4 trains of pulses.

  17. Computational Models based onHebb’s rule Many computational systems for modeling incorporate versions of Hebb’s rule. • Winner-Take-All: • Units compete to learn, or update their weights. • The processing element with the largest output is declared the winner • Lateral inhibition of its competitors. • Recruitment Learning • Learning Triangle Nodes • LTP in Episodic Memory Formation

  18. A possible computational interpretation of Hebb’s rule j wij i How often when unit j was firing, was unit i also firing? Wij = number of times both units i and j were firing ------------------------------------------------------ number of times unit j was firing

  19. Extensions of the basic idea • Bienenstock, Cooper, Munro Model (BCM model). • Basic extensions: Threshold and Decay. • Synaptic weight change proportional to the post-synaptic activation as long as the activation is above threshold • Less than threshold decreases w • Over threshold increases w • Absence of input stimulus causes the postsynaptic potential to decrease (decay) over time. • Other models (Oja’s rule) improve on this in various ways to make the rule more stable (weights in the range 0 to 1) • Many different types of networks including Hopfield networks and Boltzman machines can be trained using versions of Hebb’s rule

  20. Winner take all networks (WTA) • Often use lateral inhibition • Weights are trained using a variant of Hebb’s rule. • Useful in pruning connections • such as in axon guidance

  21. WTA: Stimulus ‘at’ is presented 1 2 a t o

  22. Competition starts at category level 1 2 a t o

  23. Competition resolves 1 2 a t o

  24. Hebbian learning takes place 1 2 a t o Category node 2 now represents ‘at’

  25. Presenting ‘to’ leads to activation of category node 1 1 2 a t o

  26. Presenting ‘to’ leads to activation of category node 1 1 2 a t o

  27. Presenting ‘to’ leads to activation of category node 1 1 2 a t o

  28. Presenting ‘to’ leads to activation of category node 1 1 2 a t o

  29. Category 1 is established through Hebbian learning as well 1 2 a t o Category node 1 now represents ‘to’

  30. Connectionist Model of Word Recognition (Rumelhart and McClelland)

  31. Recruiting connections • Given that LTP involves synaptic strength changes and Hebb’s rule involves coincident-activation based strengthening of connections • How can connections between two nodes be recruited using Hebbs’s rule?

  32. X Y

  33. X Y

  34. Finding a Connection in Random Networks For Networks with N nodes and branching factor, there is a high probability of finding good links. (Valiant 1995)

  35. Recruiting a Connection in Random Networks • Informal Algorithm • Activate the two nodes to be linked • Have nodes with double activation strengthen their active synapses (Hebb) • There is evidence for a “now print” signal based on LTP (episodic memory)

  36. Triangle nodes and recruitment Posture Push Palm

  37. Recruiting triangle nodes WTA TRIANGLE NETWORK CONCEPT UNITS A L B recruited C K free E G F

  38. Has-color Has-shape Green Round

  39. Has-color Has-shape GREEN ROUND

  40. tastebud tastes rotten eats food gets sick drinks water Hebb’s rule is insufficient • should you “punish” all the connections?

  41. So what to do? • Reinforcement Learning • Use the reward given by the environment • For every situation, based on experience learn which action(s) to take such that on average you maximize total expected reward from that situation. • There is now a biological story for reinforcement learning (later lectures).

  42. Models of Learning • Hebbian ~ coincidence • Recruitment ~ one trial • Next Lecture: Supervised ~ correction (backprop) • Reinforcement ~ delayed reward • Unsupervised ~ similarity

  43. Constraints on Connectionist Models 100 Step Rule Human reaction times ~ 100 milliseconds Neural signaling time ~ 1 millisecond Simple messages between neurons Long connections are rare No new connections during learning Developmentally plausible

  44. 5levels of Neural Theory of Language Pyscholinguistic experiments Spatial Relation Motor Control Metaphor Grammar Cognition and Language Computation Structured Connectionism abstraction Neural Net and learning SHRUTI Triangle Nodes Computational Neurobiology Biology Neural Development Quiz Midterm Finals

More Related