1 / 48

Unsupervised learning The Hebb rule – Neurons that fire together wire together. PCA

Unsupervised learning The Hebb rule – Neurons that fire together wire together. PCA RF development with PCA. Classical Conditioning and Hebb’s rule. Ear. A. Nose. B. Tongue. “When an axon in cell A is near enough to excite cell B and

paige
Download Presentation

Unsupervised learning The Hebb rule – Neurons that fire together wire together. PCA

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Unsupervised learning • The Hebb rule – Neurons that fire together wire together. • PCA • RF development with PCA

  2. Classical Conditioning and Hebb’s rule Ear A Nose B Tongue “When an axon in cell A is near enough to excite cell B and repeatedly and persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that A’s efficacy in firing B is increased” D. O. Hebb (1949)

  3. The generalized Hebb rule: where xiare the inputs and y the output is assumed linear: Results in 2D

  4. Example of Hebb in 2D w (Note: here inputs have a mean of zero)

  5. On the board: • Solve simple linear first order ODE • Fixed points and their stability for non linear ODE.

  6. In the simplest case, the change in synaptic weight w is: where xare input vectors and y is the neural response. Assume for simplicity a linear neuron: So we get: Now take an average with respect to the distribution of inputs, get:

  7. If a small change Δw occurs over a short time Δt then: (in matrix notation) If <x>=0 , Q is the covariance function. What is then the solution of this simple first order linear ODE ? (Show on board)

  8. Mathematics of the generalized Hebb rule The change in synaptic weight w is: where x are input vectors and y is the neural response. Assume for simplicity a linear neuron: So we get:

  9. Taking an average of the the distribution of inputs And using and We obtain

  10. In matrix form Where J is a matrix of ones, e is a unit vector in direction (1,1,1 … 1), and or Where

  11. The equation therefore has the form If k1 is not zero, this has a fixed point, however it is usually not stable. If k1=0 then have:

  12. The Hebb rule is unstable – how can it be stabilized while preserving its properties? The stabilized Hebb (Oja) rule. Normalize Where |w(t)|=1 Appoximate to first order in Δw: (show on board) Now insert Get:

  13. } y Therefore The Oja rule therefore has the form:

  14. Average In matrix form:

  15. Using this rule the weight vector converges to • the eigen-vector of Q with the highest eigen-value. It is often called a principal component or PCA rule. • The exact dynamics of the Oja rule have been solved by Wyatt and Elfaldel 1995 • Variants of networks that extract several principal components have been proposed (e.g: Sanger 1989)

  16. Therefore a stabilized Hebb (Oja neuron) carries out Eigen-vector, or principal component analysis (PCA).

  17. Using this rule the weight vector converges to the eigen-vector of Q with the highest eigen-value. It is often called a principal component or PCA rule. Another way to look at this: Where for the Oja rule: At the f.p: where So the f.p is an eigen-vector of Q. The condition means that w is normalized. Why? Could there be other choices for β?

  18. Show that the Oja rule converges to the state |w^2|=1 The Oja rule in matrix form: What is Bonus question for H.W: The equivalence above, why does it prove convergence to normalization.

  19. Show that the f.p of the Oja rule is such that the largest eigen-vector with the largest eigen-value (PC) is stable while others are not (from HKP – pg 202). Start with: Assume w=ua+εub where ua and ub are eigen-vectors with eigen-values λa,b

  20. Get: (show on board) Therefore: That is stable only when λa> λb for every b.

  21. Finding multiple principal components – the Sanger algorithm. Subtract projection onto accounted for subspace (Grahm-Schimdtt) Standard Oja Normalization

  22. Homework 1: (due in 1/28) Implement a simple Hebb neuron with random 2D input, tilted at an angle, θ=30o with variances 1 and 3 and mean 0. Show the synaptic weight evolution. (200 patterns at least) 1b) Calculate the correlation matrix of the input data. Find the eigen-values, eigen-vectors of this matrix. Compare to 1a. 1c) Repeat 1a for an Oja neuron, compare to 1b. + bonus question above (another 25 pt)

  23. What did we learn up to here?

  24. light electrical signals Visual Cortex Receptive fields are: • Binocular • Orientation Selective Visual Pathway Area 17 LGN Receptive fields are: • Monocular • Radially Symmetric Retina

  25. Left Right Tuning curves Response (spikes/sec) 0 90 180 270 360 Right Left

  26. Binocular Deprivation Normal Orientation Selectivity Adult Response (spikes/sec) Response (spikes/sec) Adult angle angle Eye-opening Eye-opening

  27. Monocular Deprivation Normal Left Right Right Left Response (spikes/sec) angle angle 20 30 % of cells 15 10 1 2 3 4 5 6 7 1 2 3 4 5 6 7 Rittenhouse et. al. group group

  28. First use Hebb/PCA with toy examples then used with more realistic examples

  29. Aim get selective neurons using a Hebb/PCA rule Simple example: r r r r

  30. Why? The eigen-value equation has the form: Q can be rewritten in the equivalent form: And a possible solution can be written as the sum:

  31. Inserting, and by orthogonality get: So for l=0, λ=2, and for l=1, λ=q, for l>1 there is no solution. So either w(r)= const with λ=2 or with λ=q.

  32. Orientation selectivity from a natural environment: The Images:

  33. retinal activity image Natural Images, Noise, and Learning • present patches • update weights Retina • Patches from retinal activity image LGN • Patches from noise Cortex

  34. Raw images: (fig 5.8)

  35. Preprocessed images: (fig 5.9)

  36. Monocular Deprivation Normal Left Right Right Left Response (spikes/sec) angle angle 20 30 % of cells 15 10 1 2 3 4 5 6 7 1 2 3 4 5 6 7 Rittenhouse et. al. group group

  37. Binocularity – simple examples. Q is a 2-eye correlation function. What is the solution of the eigen-value equation:

  38. In a higher dimensional case: Qll, Qlr etc. are now matrixes. And Qlr=Qrl. The eigen-vectors now have the form

  39. In a simpler case This implies Qll=Qrr, that is eyes are equivalent. And the cross eye correlation is a scaled version of the one eye correlation. If: then: with

  40. Positive correlations (η=0.2) Hebb with lower saturation at 0 Negative correlations (η=-0.2)

  41. Lets now assume that Q is as above for the 1D selectivity example.

  42. With 2D space included

  43. 2 partially overlapping eyes using natural images

  44. Right Eye Left Eye 100 50 0 100 Right Synapses 50 Left Synapses 0 100 50 0 100 50 0 PCA Left Right Orientation selectivity and Ocular Dominance No. of Cells 1 3 5 Bin

  45. What did we learned today?

  46. The Hebb rule is unstable – how can it be stabilized while preserving its properties? The stabilized Hebb (Oja) rule. This is has a fixed point at: Where: The only stable fixed point is for λmax

More Related