the helmholtz machine p dayan, ge hinton, rm neal, rs zemel

21
The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS Zemel Computational Modeling of Intelligence 11.04.22.(Fri) Summarized by Joon Shik Kim

Upload: kyran

Post on 24-Feb-2016

69 views

Category:

Documents


0 download

DESCRIPTION

The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS Zemel. Computational Modeling of Intelligence 11.04.22.(Fri) Summarized by Joon Shik Kim. Abstract. Discovering the structure inherent in a set of patterns, is a fundamental aim of statistical inference or learning. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Helmholtz MachineP Dayan, GE Hinton, RM Neal, RS Zemel

Computational Modeling of Intelli-gence

11.04.22.(Fri)Summarized by Joon Shik Kim

Page 2: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

Abstract• Discovering the structure inherent in a set of

patterns, is a fundamental aim of statistical in-ference or learning.

• One fruitful approach is to build a parametrized stochastic generative model.

• Each pattern can be generated in exponentially many ways.

• We describe a way of fitnessing this combinato-rial explosion by maximizing an easily com-puted lower bound on the probability of the ob-servation.

Page 3: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

Introduction (1/3)• Following Helmholtz, we view the

human perceptual system as a statis-tical inference engine.

• Inference engine’s function is to infer the probable causes of sensory input.

• A recognition model is used to infer a probability distribution over the un-derlying causes from the sensory in-put.

Page 4: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

Introduction (2/3)• A separate generative model is used

to train the recognition model.

Figure 1: Shift patterns.

Page 5: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

Introduction (3/3)• First the direction of the shift was chosen.• Each pixel in the bottom row was turned on

(white) with a probability of 0.2.• Shifted pixel in the top row and the copies of

these in the middle rows were made.• If we treat the top two rows as a left retina and

the bottom two rows as a right retina, detecting the direction of the shift resembles the task of extracting depth from simple stereo images.

• Extra redundancy facilitates the search for the correct generative model.

Page 6: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Recognition Distribution (1/6)

• The log probability of generating a particular example, d, from a model with parameter θ is

where α are explanations. We call each possible set of causes an explanation of the pattern.

Page 7: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Recognition Distribution (2/6)

• We define the energy of explanation α to be

• The posterior distribution of an ex-planation given d and θ is

Page 8: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Recognition Distribution (3/6)• The Helmholtz free energy is

• Any probability distribution over the explanations will have at least as high a free energy as the Boltzmann distribution.

• The log probability of the data can be written as

where F is the free energy based on the incorrect or nonequilibrium posterior Q.

Page 9: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Recognition Distribution (4/6)• The last term in above equation is the Kull-

back-Leibler divergence between Q(d) and the posterior distribution P(θ,d).

• Distribution Q is produced by a separate recognition model that has its own parame-ters, Ф.

• These parameters are optimized at the same time as the parameters of the gener-ative model, θ, to maximize the overall fit function .

Page 10: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Recognition Distribution (5/6)

Figure 2: Graphical view of our approximation.

Page 11: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Recognition Distribution (6/6)

Page 12: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Deterministic Helmholtz Ma-chine (1/5)

• Helmholtz machine is a connectionist system with multiple layers of neuron-like binary stochastic processing units connected hierarchically by two sets of weights.

• Top-down connections θ implement the generative model.

• Bottom-up connections Ф implement the recognition model.

Page 13: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Deterministic Helmholtz Ma-chine (2/5)

• The key simplifying assumption is that the recognition distribution for a particular exam-ple d, Q(Ф,d), is factorial (separable) in each layer.

• Let’s assume there are h stochastic binary units in a layer l.

• Q(Ф,d) makes the assumption that the actual activity of any one unit in layer l is indepen-dent of the activities of all the other units in that layer, so the recognition model needs only specify h probabilities rather than 2h-1.

Page 14: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Deterministic Helmholtz Ma-chine (3/5)

where .

Page 15: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Deterministic Helmholtz Ma-chine (4/5)

Page 16: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Deterministic Helmholtz Ma-chine (5/5)

• One could perform stochastic gradi-ent ascent in the negative free en-ergy across all the data with the above equation and a form of REIN-FORCEMENT algorithm.

Page 17: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Wake-Sleep Algorithm (1/3)

• Borrowing an idea from the Boltzmann machine, we get the wake-sleep algo-rithm, which is a very simple learning scheme for layered networks of stochastic binary units.

• During the wake phase, data d from the world are presented at the lowest layer and binary activations of units at succes-sively higher layers are picked according to the recognition probabilities.

Page 18: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Wake-Sleep Algorithm (2/3)

• The top-down generative weights from layer l+1 to l are then altered to reduce the Kullback-Leibler divergence between actual activations and the generated probabilities.

• In the sleep phase, the recognition weights are turned off and the top-down weights are used to activate the units.

• The network thus generates a random in-stance from its generative model.

Page 19: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

The Wake-Sleep Algorithm (3/3)

• Since it has generated the instance, it knows the true underlying causes, and therefore has the available tar-get values for the hidden units that are required to train the bottom-up weights.

• Unfortunately, there is no single cost function that is reduced by these two procedures.

Page 20: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

Discussion (1/2)• Instead of using a fixed independent prior

distribution for each of the hidden units in a layer, the Helmholtz machine makes this prior more flexible by deriving it from the bottom-up activities of units.

• The old idea of analysis-by-synthesis as-sumes that the cortex contains a genera-tive model of the world and that recogni-tion involves inverting the generative model in real time.

Page 21: The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS  Zemel

Discussion (2/2)• An interesting first step toward inter-

action within layers would be to or-ganize their units into small clusters with local excitation and longer-range inhibition, as is seen in the columnar structure of the brain.