computation in sensory processing · • adrian (1920s) • first recordings ... •learning of...

97
Spikes and Computation in Sensory Processing Simon Thorpe CerCo (Brain and Cognition Research Center) & SpikeNet Technology SARL, Toulouse, France

Upload: others

Post on 18-Oct-2020

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Spikes and Computation in

Sensory Processing Simon Thorpe

CerCo (Brain and Cognition Research Center) &

SpikeNet Technology SARL,

Toulouse, France

Page 2: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

• 1974-77Physiology and Psychology (Oxford)• 1977-81 Doctorate with Edmund Rolls (Oxford)

• Recording neurons in orbitofrontal cortex, striatum, parietal cortex, lateral hypothalamus, amygdala, hippocampus, supplementary motor area (!!)

• 1982-83Post-doc with Max Cynader (Halifax, Canada)• Attempting to test Hebb's hypothesis in kitten visual cortex

• 1983-93Paris with Michel Imbert• Testing Hebb's Hypothesis (with Yves Frégnac & Elie Bienenstock)• Dynamics of responses in monkey V1 (with Simona Celebrini & Yves Trotter)

• 1993-nowCERCO Toulouse• Work on ultra-rapid scene processing

• 1999-now SpikeNet • Spike-based image processing• Bioinspired vision

Simon Thorpe

Page 3: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Overview

Part 1

• Ultrarapid visual categorization

• Biological vs Computer Vision

• Temporal Constraints

• Coding with Spikes

• Convolutional Neural Networks in 1999

• SpikeNet

• The current state of the art in Convolutional Neural Networks

• Supervision & GoogleNet

Part 2

• Biologically inspired learning

• Spike-Time Dependent Plasticity (STDP)

• Applications in Vision

• Applications in Audition

• Development of Neural Selectivity

• Memories that can last a lifetime

• Grandmother Cells

• Neocortical Dark Matter

Page 4: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Behavioural Reaction Times

Targets

Distractors

Difference

Event Related Potentials

Scene Processing in 150 ms

Ultra-Rapid Visual Processing

Page 5: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

•Saccades towards faces

•Latency 100ms!

Crouzet,  Kirchner  &  Thorpe,  2010

Ultra-Rapid Visual Processing

Page 6: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

RSVP at 10 fps

Ultra-Rapid Visual Processing

Page 7: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Activating Memories

• “That Dalmation picture that I saw in Psych 101”

• How does the brain recognize visual objects and scenes?

• Can a machine be built that can do the same thing?

Some Questions

Page 8: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Biological and Computer VisionCommon Problems

• Identification and Categorising objects and events in complex dynamically changing natural scenes• As fast as possible• As reliably as possible• Using the most energy efficient hardware possible• Using the smallest size and weight footprint

Common Solutions?• David Marr (1982) “Vision : A computational

investigation into the Human Representation and Processing of Visual Information”

• Recent years - is there convergence?

Page 9: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Hardware ConstraintsElectronics • Nvidia GeForce GTX Titan

• 4.5 TeraFlops!

• 2668 cores

• 7.1 billion transistors

• 288 Gbytes/sec Memory bus

• $999

Brain• 1 KHz

• 86 billion processors

• 1-2 m/s conduction velocity

• 20 watts

QuestionsAre we going to be able to implement brain style computing with conventional computing?

ResponseIt depends if we can understand how the brain computes

How many teraflops does the brain need?

How much memory bandwidth?

Page 10: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Classic Neural Computing

• To simulate the visual system

• 4 billion neurons

• 10000 connections each

• Update at 1 kHz

• 40 Petaflops

Page 11: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

What’s missing?

• Real brains use spikes• Simulation

• 16.7 million neurons

• 21 billion synapses

• The Brain• 86 billion neurons

• 16 billion in the cortex

• 100 trillion synapses

• Spikes• Firing rate

• 0-200 Hz

• Spontaneous activity

• 1-2 Hz?

Page 12: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Classes of Brain Simulation

• Connectionist Simulators (non-spiking)

• Backpropagation, ART, Kohonen Maps, Time Delay Networks,

• Computational Neuroscience Simulators

• Up to 30 000 compartments

• Lots of channel kinetics etc.

• Spike-Based Simulators

• Software

• Hardware

Page 13: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

The Human Brain Project

• European Flagship Proposal

• Three simulation approaches

• Blue Brain

• Henry Markram (Lausanne, Switzerland)

• Analog Chips

• Karl-Heinz Meyer (Heidelburg, Germany)

• SpiNNaker

• Steve Furber (Manchester, UK)

Page 14: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

SpiNNaker Project• A simulation system for Spiking Neural Networks

• Prof. Steve Furber, Computer Science, University of Manchester

• 18 ARM968 cores • each with 64 Kbytes of Data and 32 Kbytes of Instructions

• 128 MBytes of shared memory• 48 chips on a board• 18 boards in a 19” frame

A billion neurons in real time

Page 15: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

IBM TrueNorth Chip

• What can be computed?

SCIENCE

Page 16: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

• Jerry Feldman’s 100-step limit

• High level decisions in about 0.5 seconds

• Interval between spikes around 5 ms

• No more than 100 (massively parallel) computational steps

• Development of connectionist and PDP modelling

• Surely, more detail can help

Temporal Constraints – Early 1980s

Page 17: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Temporal Contraints - 1989

Argument• Roughly 10 layers• 10 ms per layer• Firing rates 0-100 Hz

Face selectivity at 100 ms

Food selectivity at 150 ms

Therefore• Mainly feedforward (?!) • One spike per neuron (?!)

T

T

T

TT

TRetina

LGN

V1

V2

V4

IT

Page 18: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Temporal Constraints 1988-1989Inferotemporal Cortex :Face selectivity at 100 ms

Perrett, Caan & Rolls, 1982

See also : Jerry Feldman & John TsotsosLeonard Uhr (1987)

Page 19: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

How can you code with just one spike per neuron?

• Feedforward processing

• Only a few milliseconds per processing step

• One spike per neuron

• Very sparse coding

• Processing without context based help

Ultra-Rapid Visual Processing

Page 20: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Sensory Coding with Spikes

• Adrian (1920s)• First recordings from sensory fibres

• Hubel & Wiesel (1962)• Orientation selectivity in striate cortec

Page 21: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Sensory Coding with Spikes

• Wurtz & Goldberg (1972, 1976)

Raster Display Post Stimulus Time Histogram

Assumption : Firing rate is enough to describe the response

Page 22: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

The Classic View

2

4

1

2

3

0,213

0,432

0,112

0,238

0,309

•Spikes don't really matter

•Neurons send floating point numbers

•The floating point numbers are transformed into spikes trains using a Poisson process

•God plays dice with spike generation0,375

Page 23: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Temporal Coding Option

• Spikes do really matter

• The temporal patterning of spikes across neurons is critical for computation

• Synchrony• Repeating patterns• etc

• The apparent noise in spiking is unexplained variation

Page 24: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Simon Thorpe's Version

Weak StimulusMedium StimulusStrong

Stimulus

Time

Threshold

•Ordering of spikes is critical•The most activated neurons fire first•Temporal coding is used even for stimuli that are not temporally structured•Computation theoretically possible even when each neuron emits one spike

Page 25: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Sensory Coding with Spikes• Adrian (1920s)

• First recordings from sensory fibres

Low  intensity

High  intensity

Higher  maintained  firing

Higher  peak  firing

Shorter  Latency!

Page 26: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Coding in the Optic Nerve

1,000,000 fibres

Page 27: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Intensity

n

n

n

n

n

n

n

n

nn

n

n

n

n

n

n

Coding in the Optic Nerve

Page 28: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Intensity

n

n

n

n

n

n

n

n

nn

n

n

n

n

n

n

A mini retina32 x 32 pixels

Coding in the Optic Nerve

Page 29: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Coding with Spike Ordering

Less than 1% of cells need to fire for recognition!

Example• A toy retina

Page 30: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Early Studies

• Face identification directly from the output of oriented filters

Page 31: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Early Studies• Virtually all the faces correctly

identified

• Very robust to low contrast

• Very robust to noise

Page 32: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

SpikeNet Technology

• Created in 1999

• Simon Thorpe, Rufin VanRullen & Arnaud Delorme

• Currently 12 employees

Page 33: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

SpikeNet - Invariance

• Luminance

• Contrast

• Blurring

• Noise

Page 34: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

SpikeNet - Invariance

• Size

• Rotation • Perspective

Page 35: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

SpikeNet - Invariance

• 3D Rotation

• Identity

•What is the current state of the art?

Page 36: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Biological vs Computer HardwareComputer

• Nvidia GeForce GTX Titan

• 4.5 TeraFlops!

• 2668 cores

• 7.1 billion transistors

• 288 Gbytes/sec Memory bus

• 200 watts

• $999

• 86 billion neurons

• 16 billion in the cortex

• 4 billion in the visual system

• 1 KHz

• 1-2 m/s conduction velocity

• 20 wattsIs that enough to reproduce human performance?

Brain

Page 37: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

The ImageNet Challenge

• 10,000,000 training images

• 10,000+ labels

• Systems tested on new images, with 1000 possible labels

• ECCV 2012 Firenze

• The state of the art was beaten by a “simple” feedforward convolutional neural network trained with Back-Propagation

Page 38: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

SuperVision

• 650,000 “neurons”

• 60,000,000 parameters

• 630,000,000 “synapses”

Output Layer

Convolutional Layers

Fully Connected Layers

253440 186624 64896 64896 43264

Page 39: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Animals

SuperVision

Page 40: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

And then….• Geoff Hinton and his two

students launched a start-up (DNNresearch)

• bought by Google…

• Yann LeCun, a pioneer of feed-forward convolutional networks since the end of the 1980s

• Hired by Facebook…

Page 41: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Supervision vs Primate Vision

Convergent Evolution!

Page 42: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

? ? ? ?

Coding at higher levels

Page 43: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Basic Methodology

•Computer models• 57 parameters• 1-3 layers

•Neural Activity• Parallel Multielectrode Recordings • V4 and IT• Activity (70-170 ms)

•Human performance• Crowd sourced data• Amazon Mechanical Turk• 104 subjects

Page 44: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Comparison of Models

• HMO - Hierarchical Modular Optimisation

• 4 layer CNN with 1250 output units

• Close match between models, humans and neural signals

Page 45: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Human vs Machine

Page 46: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

• Sorry, Andrej!

• What about faces?

Feb  2015

Human vs Machine

Page 47: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

A range of architectures• SpikeNet (1999)

• Identification immediately after V1

• Supervision (2012)• 7 layers

• GoogleNet (2014)• 30 layers

• 12 times less parameters

Page 48: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

The Surprises• Human levels of performance are possible despite

• No feedback• No horizontal connections• No top-down control• No need for context• No dendrites• Complex channel dynamics• Only positive and negative weights• No dynamics• No memory• No oscillations• No “binding”• No attention!!!• No Spikes!!!

Page 49: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Question : Is Spike Timing Precise Enough?

Fred Rieke, David Warland, Rob de Ruyter van Steveninck and William Bialek (1999)

Increasing evidence that the timing of spikes carries a great deal of information

Page 50: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

First Spike Coding : Evidence

Coding motion direction

Coding stimulus shape

Page 51: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Experimental proof in the visual system!

Salamander Retinal Ganglion cellsInformation in latency vs Information in the spike count

Page 52: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Overview

Part 1

• Ultrarapid visual categorization

• Biological vs Computer Vision

• Temporal Constraints

• Coding with Spikes

• Convolutional Neural Networks in 1999

• SpikeNet

• The current state of the art in Convolutional Neural Networks

• Supervision & GoogleNet

Part 2

• Biologically inspired learning

• Spike-Time Dependent Plasticity (STDP)

• Applications in Vision

• Applications in Audition

• Development of Neural Selectivity

• Memories that can last a lifetime

• Grandmother Cells

• Neocortical Dark Matter

Page 53: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

The Learning Problem

•SuperVision uses Backpropagation

•Billions of training examples with labeled data

• 1 ms per image!• 100 images in 100 ms

•Totally unbiological

Spike-Based Processing

Page 54: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Song, Miller & Abbott, 2000

• Synapses that fire before the target neuron get strengthened

• Synapses that fire after the target neuron get weakened

Spike Time Dependent Plasticity (STDP)

Simplified STDP rule• All synapses are depressed after each output neuron spike

• Except those that fire in just before

• Natural Consequence• High synaptic weights concentrate on early firing inputs

Page 55: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Threshold 3

Short time constant

12 inputs

Initial synaptic weight = 0.25

Threshold = 3.0

All inputs must fire to reach threshold

Page 56: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Learning of a repeating pattern

Threshold 3

Short time constant

Presentation 1

All synapses activated before the output

neuron

STDP reinforces all synapses

Page 57: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Learning of a repeating pattern

Threshold 3

Short time constant

Presentation 2

9 synapses reinforced

3 depressed

Page 58: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Learning of a repeating pattern

Threshold 3

Short time constant

Presentation 3

6 synapses reinforced

6 depressed

Page 59: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Learning of a repeating pattern

Threshold 3

Short time constant

Presentation 4

3 synapses reinforced

9 depressed

Page 60: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Learning of a repeating pattern

Threshold 3

Short time constant

Presentation 5

Page 61: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Learning of a repeating pattern

Threshold 3

Short time constant

Presentation 6

3 fully potentiated synapses

9 set to zero

Page 62: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Learning of a repeating pattern

Threshold 3

Short time constant

In six presentations STDP has found the first three inputs that

fire duing the repeating pattern

Page 63: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Multiple Units

Mutual inhibition

Competitive Learning Mechanism

Page 64: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Multiple Units

Page 65: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Multiple Units

Page 66: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Multiple Units

Page 67: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Multiple Units

Neurons can detect patterns embedded in continous activity

Page 68: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Neurons can detect patterns embedded in continous activity

Multiple Units

Add a second layer to detect combinations of combinations

Can this be scaled up beyond a toy demo?

Page 69: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Learning Spike Sequences with STDP

Page 70: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

• Initial State

• DuringLearning

• After Learning

Learning Spike Sequences with STDP

Page 71: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Learning Faces with STDP

Page 72: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Learning Cars with STDP

Tobi Delbruck’s Spiking Retina

STDP Rule

Page 73: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

12 seconds 30 seconds 90 seconds

• The system has learned to recognize cars

• No supervision!

“STDP” Rule

Tobi Delbruck’s Spiking Retina

Learning Cars with STDP

Page 74: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Building a complete visual system

1920

1080

• Add more neurons

• Add more layers of processing

• Increase the input resolution

• Use more sophisticated preprocessing

• Mexican hat convolutions

• Different types of input signal

Page 75: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Auditory Noise Learning in Humans

•Learning of random noise patterns

•Roughly 10 repetitions are sufficient!

•Learning appears to be all-or-none

•Lasts for weeks at least

Page 76: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Auditory Noise Learning with STDP

Olivier Bichler, Thesis

Modified STDP rule•Post synaptic spike - depresses all synapses except those activated recently

Page 77: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Auditory Noise Learning with STDP

Page 78: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Auditory Noise Learning -ERP data

Perceptual  learning  of  acoustic  noise  generates  memory-­‐evoked  potentials  Thomas  Andrillon,        Sid  Kouider,  Trevor  Agus    &  Daniel  Pressnitzer  (Currrent  Biology,  accepted)

Selective Brain Responses to Meaningless noise in just 5 presentations

Page 79: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Selectivity MechanismsThreshold

Page 80: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Controlling Sparsity

• With Temporal Coding

• Easy to control the percentage of neurons that fire

• 1%, 2%, 5%, 10%

Page 81: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

12345678910

Controlling Sparsity

• With Temporal Coding

• Easy to control the percentage of neurons that fire

• 1%, 2%, 5%, 10%

Page 82: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

1,00E+00&

1,00E+02&

1,00E+04&

1,00E+06&

1,00E+08&

1,00E+10&

1,00E+12&

1,00E+14&

1,00E+16&

1,00E+18&

1,00E+20&

1,00E+22&

1,00E+24&

1,00E+26&

1,00E+28&

1,00E+30&

0& 10& 20& 30& 40& 50& 60& 70& 80& 90& 100&

Combina(ons*with*M*=*100*

12345678910

N of M coding

• Steve Furber

• With 100 neurons• 2100 binary patterns• ~1030

• What about controlling the proportion of active neurons?

1,00E+024,95E+031,62E+053,92E+067,53E+071,19E+091,60E+101,86E+111,90E+121,73E+13

Page 83: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Generating Selectivity• Suppose that the neuron has 10 synapses

• Suppose that the percentage of active inputs is fixed at 10%

• What is the likelihood of having a given number of synapses active?

• With a threshold of 4 the neuron would only have a 1% chance of firing with a random input

• With a threshold of 5, the probability drops to 0.1%

Page 84: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Interim Conclusion - Learning

• Error back-propagation is an unrealistic way to determine the selectivity of neurons in the visual system

• But simple spike-based learning mechanisms could allow the system to generate the appropriate selectivity in an unsupervised way

• Can we build a visual system without label based training??

Page 85: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Memory Mechanisms in Man & Machine

1. Humans can recognise visual and auditory stimuli that they have not experienced for decades.

2. Recognition after very long delays is possible without ever reactivating the memory trace in the intervening period.

3. These very long term memories require an initial memorisation phase, during which memory strength increases roughly linearly with the number of presentations

4. A few tens of presentations can be enough to form a memory that can last a lifetime.

5. Attention-related oscillatory brain activity can help store memories efficiently and rapidly

10 provocative claims

Page 86: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Memory Mechanisms in Man & Machine

6. Storing such very long-term memories involves the creation of highly selective "Grandmother Cells" that only fire if the original training stimulus is experienced again.

7. The neocortex contains large numbers of totally silent cells ("Neocortical Dark Matter") that constitute the long-term memory store.

8. Grandmother Cells can be produced using simple spiking neural network models with Spike-Time Dependent Plasticity (STDP) and competitive inhibitory lateral connections.

9. This selectivity only requires binary synaptic weights that are either "on" or "off", greatly simplifying the problem of maintaining the memory over long periods.

10. Artificial systems using memristor-like devices can implement the same principles, allowing the development of powerful new processing architectures that could replace conventional computing hardware.

10 provocative claims

Page 87: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

STDP and Spiking Neurons

• “Intelligent” sensory processing with small numbers of neurons connected to spiking sensory devices• 2 layers of neurons connected to a spiking retina

• Unsupervised learning of car counting

• A single neuron connected to a spiking cochlear

• Unsupervised learning of auditory noise patterns

• What would happen with 16 billion neurons and multiple sensory inputs?

16 billion neurons

2 million optic nerve fibres

60,000 auditory nerve fibres

Page 88: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Grandmother Cells?

• Jerzy Konorski (1967)

• “gnostic neurons”

• Horace Barlow (1972)

• “cardinal cells”

• Alberta Gilinsky (1984)

• “cognons”

Page 89: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

A Jennifer Anniston Cell !

Grandmother Cells in Man?

• Halle Berry• The Taj Mahal• Bill Clinton• Saddam Hussein• The Simpsons• The patient’s brother• Members of the research

team• ...

Page 90: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Temporal Lobe Recordings in Humans

“Josh Brolin” “Marilyn Munroe”

Page 91: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

• The hit rate (0.54%) is too high

• Assume 109 neurons in the hippocampus

• Assume 10-30 000 identifiable objects (Irv Biederman)

• Each neuron should respond to 100-150 different objects

• Number of identifiable objects probably much higher : 105 - 106

• Rafi Malach : “Totem Pole Cells”

• Alternative Hypothesis

• The hippocampus keeps track of a few thousand objects and events that have been experiences recently

Grandmother Cells in Man?

Page 92: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

How to make a Grandmother Cell

• STPD allows neurons to become rapidly selective to repeated stimuli• This can happen in a few tens of presentations• Oscillatory activity could allow this to happen in a single presentation

• Controlling the percentage of active inputs guarantees that the neuron will not fire to random inputs

• Setting a high threshold means that the neuron will never fire unless the stimulus used for training was presented again

• It could maintain its selectivity for several decades

• Cortical Dark Matter?

Page 93: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Dark Matter in Cortex?• Question

• What is the true distribution of firing rates in the cortex?

• Are there neurons that never fire for very long periods of time?

• Problem• Nearly all single unit neurophysiological studies are biased towards neurons that

have spontaneous activity

• The number of neurons recorded seems far lower than would be expected• Only 5-15 neurons recorded per descent (out of hundreds)

• Possible methods• Implanted arrays

• 2-photon optical imaging

• Patch clamp recording

Page 94: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

• Roughly 50% of output neurons have firing rates below 0.1 Hz!• Maybe the output neurons are a special class??

• Dark matter in cortex??

Very low firing rate neurons

Page 95: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Invisible Cells?

Page 96: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Overview

Part 1

• Ultrarapid visual categorization

• Biological vs Computer Vision

• Temporal Constraints

• Coding with Spikes

• Convolutional Neural Networks in 1999

• SpikeNet

• The current state of the art in Convolutional Neural Networks

• Supervision & GoogleNet

Part 2

• Biologically inspired learning

• Spike-Time Dependent Plasticity (STDP)

• Applications in Vision

• Applications in Audition

• Development of Neural Selectivity

• Memories that can last a lifetime

• Grandmother Cells

• Neocortical Dark Matter

Page 97: Computation in Sensory Processing · • Adrian (1920s) • First recordings ... •Learning of random noise patterns •Roughly 10 repetitions are sufficient! •Learning appears

Final Conclusions

• Temporal Constraints on Vision

• Feedforward Neural Networks

• SpikeNet

• Supervision & GoogleNet

• Importance of Spikes

• Ultrafast coding with one spike per neuron

• Very efficient STDP schemes for unsupervised learning

• Biological Vision has been and will continue to be a source of inspiration for technology