90 likes | 214 Vues
This project explores artificial life generation through recurrent neural networks (RNNs) and Hebbian learning. By deriving RNNs from genomes, organisms are created with a 360-degree visual system. The encoding methods, including graph grammar and cellular encoding, facilitate the generation of adjacency matrices that represent neural network structures. The organisms navigate a wrap-around grid world where they can see and consume food, utilizing their RNN to determine actions. While foundational concepts are implemented, several aspects and extensions of the model remain to be developed.
E N D
Artificial Life Using Recurrent Neural Networks and Hebbian Learning Michael Tauraso CS 152
Organisms • Behavior decided by a Recurrent Neural Network (RNN) • RNN derived from a genome • Organisms have a 360 degree visual system
Gene Encoding • Direct Encoding (boring) • Graph Grammar (interesting) • Cellular Encoding (Interesting) • Functional Block Problem
Graph Grammar • Goal is to generate an adjacency matrix • Each non-terminal generates a 2x2 matrix • Each 2x2 matrix contains either numbers or other non-terminals. • The start symbol is used to build up an adjacency matrix
Cellular Encoding • Fundamental unit of genome is a transformation. • Transformations act on a particular neuron • A sequence of transformations defines a neural network. • This preserves functional blocks better
Grid World • World is a wrap-around grid • Each grid can have 1 organism • Each grid can have food on it for the organisms to eat • Organisms have a 360 degree visual system • Organisms use an RNN to decide on their actions
Learning Method • Recurrent networks • Hebbian Learning • Clamping Outputs
Results? • Grid World difficult to implement • Two Interesting encodings remain unimplemented • Interesting extensions to the world model remain unimplemented