Artificial Neural Networks: Torsten Reil
Artificial Neural Networks: Torsten Reil
Artificial Neural Networks: Torsten Reil
Torsten Reil
torsten.reil@zoo.ox.ac.uk
Outline
• What are Neural Networks?
• Biological Neural Networks
• ANN – The basics
• Feed forward net
• Training
• Example – Voice recognition
• Applications – Feed forward nets
• Recurrency
• Elman nets
• Hopfield nets
• Central Pattern Generators
• Conclusion
What are Neural Networks?
• Applications
– As powerful problem solvers
– As biological models
Biological Neural Nets
– Experiment:
• Pigeon in Skinner box
• Present paintings of two different artists (e.g. Chagall / Van
Gogh)
• Reward for pecking when presented a particular artist (e.g. Van
Gogh)
• Pigeons were able to discriminate between Van Gogh
and Chagall with 95% accuracy (when presented with
pictures they had been trained on)
1. Neurones (nodes)
2. Synapses (weights)
• Neurone vs. Node
• Structure of a node:
• Information is distributed
1
Squashing: 0.3775
1 e 0.5
• Data is presented to the network in the form of
activations in the input layer
• Examples
– Pixel intensity (for pictures)
– Molecule concentrations (for artificial nose)
– Share prices (for stock market prediction)
• Backpropagation
– Requires training set (input / output pairs)
– Starts with small random weights
– Error is used to adjust weights (supervised learning)
Gradient descent on error landscape
• Advantages
– It works!
– Relatively fast
• Downsides
– Requires a training set
– Can be slow
– Probably not biologically realistic
• Alternatives to Backpropagation
– Hebbian learning
• Not successful in feed-forward nets
– Reinforcement learning
• Only limited success
– Artificial evolution
• More general, but can be even slower than backprop
Example: Voice Recognition
• Data
– Sources
• Steve Simpson
• David Raubenheimer
– Format
• Frequency distribution (60 bins)
• Analogy: cochlea
• Network architecture
– Feed forward network
• 60 input (one for each frequency bin)
• 6 hidden
• 2 output (0-1 for “Steve”, 1-0 for “David”)
• Presenting the data
Steve
David
• Presenting the data (untrained network)
Steve
0.43
0.26
David
0.73
0.55
• Calculate error
Steve
0.43 – 0 = 0.43
0.26 –1 = 0.74
David
0.73 – 1 = 0.27
0.55 – 0 = 0.55
• Backprop error and adjust weights
Steve
0.43 – 0 = 0.43
0.26 – 1 = 0.74
1.17
David
0.73 – 1 = 0.27
0.55 – 0 = 0.55
0.82
• Repeat process (sweep) for all training pairs
– Present data
– Calculate error
– Backpropagate error
– Adjust weights
0.01
0.99
David
0.99
0.01
• Results – Voice Recognition
• Demo
• Results – Voice Recognition (ctnd.)
– Stock-market prediction
– Pronunciation (NETtalk)
(Sejnowksi & Rosenberg, 1987)
Cluster analysis of hidden layer
FFNs as Biological Modelling Tools
• Recurrency
– Nodes connect back to other nodes or themselves
– Information flow is multidirectional
– Sense of time and memory of previous state(s)
• Task
– Elman net to predict successive words in sentences.
• Data
– Suite of sentences, e.g.
• “The boy catches the ball.”
• “The girl eats an apple.”
– Words are input one at a time
• Representation
– Binary representation for each word, e.g.
• 0-1-0-0-0 for “girl”
• Training method
– Backpropagation
• Internal representation of words
Hopfield Networks
• Sub-type of recurrent neural nets
– Fully recurrent
– Weights are symmetric
– Nodes can only be on or off
– Random updating
auto-associative or
content-addressable memory
Task: store images with resolution of 20x20 pixels
Hopfield net with 400 nodes
Memorise:
1. Present image
2. Apply Hebb rule (cells that fire together, wire together)
• Increase weight between two nodes if both have same activity, otherwise decrease
3. Go to 1
Recall:
1. Present incomplete pattern
2. Pick random node, update
3. Go to 2 until settled
DEMO
• Memories are attractors in state space
Catastrophic forgetting
Computer modelling
– E.g. lamprey swimming (Ijspeert et al., 1998)
• Evolution of Bipedal Walking (Reil & Husbands, 2001)
0.9
0.8
0.7
left hip lateral
activation
0.2
0.1
0
1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 33 35 37 39 41 43 45 47 49 51 53 55 57 59 61 63 65 67 69 71 73 75 77
time
• CPG cycles are cyclic attractors in state space
Recap – Neural Networks
• Components – biological plausibility
– Neurone / node
– Synapse / weight
• Recurrent networks
– Multidirectional flow of information
– Memory / sense of time
– Complex temporal dynamics (e.g. CPGs)
– Various training methods (Hebbian, evolution)
– Often better biological models than FFNs
Online material:
http://users.ox.ac.uk/~quee0818