1 / 78

Hopfield NNets

Hopfield NNets. N. Laskaris. Professor John Hopfield The Howard A. Prior Professor of Molecular Biology Dept. of Molecular Biology Computational Neurobiology; Biophysics Princeton University.

stash
Download Presentation

Hopfield NNets

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Hopfield NNets N. Laskaris

  2. Professor John Hopfield The Howard A. Prior Professor of Molecular Biology Dept. of Molecular Biology Computational Neurobiology; Biophysics Princeton University

  3. The physicist Hopfield showed that models of physical systems could be used to solve computational problems Such systems could be implemented in hardware by combining standard components such as capacitors and resistors.

  4. The importance of the Hopfield nets in practical application is limited due to theoretical limitations of the structure, but, in some cases, they may form interesting models.

  5. Usually employed in binary-logic tasks : e.g. pattern completion and association

  6. The concept

  7. In the beginning of 80s Hopfield published two scientific papers, which attracted much interest. (1982):‘’Neural networks and physical systems with emergent collective computational abilities’’.Proceedings of the National Academy of Sciences, pp. 2554-2558. (1984):‘’Neurons with graded response have collective computational properties like those of two-state neurons’’.Proceedings of the National Academy of Sciences, pp. 81:3088-3092 This was the starting point of the new era of neural networks, which continues today 

  8. ‘‘The dynamics of brain computation” The core question : How is one to understand the incredible effectiveness of a brain in tasks such as recognizing a particular face in a complex scene?

  9. Like all computers, a brain is a dynamical system that carries out its computations by the change of its 'state' with time. Simple models of the dynamics of neural circuits are described that have collective dynamical properties. These can be exploited in recognizing sensory patterns. Using these collective properties in processing information is effective in that it exploits thespontaneousproperties of nerve cells and circuits to produce robust computation.

  10. J. Hopfield’s quest While the brain is totally unlike modern computers, much of what it does can be described as computation. Associative memory, logic and inference, recognizing an odor or a chess position, parsing the world into objects, and generating appropriate sequences of locomotor muscle commands are all describable as computation. His research focuses on understanding how the neural circuits of the brain produce such powerful and complex computations.

  11. The simplest problem in olfaction is simply identifying a known odor. Olfaction However, olfaction allows remote sensing, and much more complex computations involving wind direction and fluctuating mixtures of odors must be described to account for the ability of homing pigeons or slugs to navigate through the use of odors. Hopfield has been studying how such computations might be performed by the known neural circuitry of the olfactory bulb and prepiriform cortex of mammals or the analogous circuits of simpler animals.

  12. Dynamical systems Any computer does its computation by its changes in internal state. In neurobiology, the change of potentials of neurons (and changes in the strengths of the synapses) with time is what performs the computations. Systems of differential equations can represent these aspects of neurobiology. He seeks to understand some aspects of neurobiological computation through studying the behavior of equations modeling the time-evolution of neural activity.

  13. Action potential computation For much of neurobiology, information is represented by the paradigm of ‘‘firing rates’’, i.e. information is represented by the rate of generation of action potential spikes, and the exact timing of these spikes is unimportant.

  14. Action potential computation Since action potentials last only about a millisecond, the use of action potential timing seems a powerful potential means of neural computation.

  15. Action potential computation There are cases, for examplethe binaural auditory determination of the location of a sound source, where information is encoded in the timing of action potentials.

  16. Speech Identifying words in natural speech is a difficult computational task which brains can easily do. They use this task as a test-bed for thinking about the computational abilities of neural networks and neuromorphic ideas

  17. Simple (e.g. binary-logic ) neurons are coupled in a system with recurrent signal flow

  18. A 2-neurons Hopfield network of continuous states characterized by 2 stable states 1st Example Contour-plot

  19. 2nd Example A 3-neurons Hopfield network of 23=8 states characterized by 2 stable states

  20. 3rd Example Wij = Wji The behavior of such a dynamical system is fully determined by the synaptic weights And can be thought of as an Energy minimization process

  21. Hopfield Nets are fully connected, symmetrically-weighted networks that extended the ideas of linear associative memories by adding cyclic connections . Note: no self-feedback !

  22. Operation of the network After the ‘teaching-stage’, in which the weights are defined, the initial state of the network is set (input pattern) and a simple recurrent rule is iterated till convergence to a stable state (output pattern) Regarding training a Hopfield net as a content-addressable memory the outer-product rule for storing patterns is used There are two main modes of operation: Synchronous vs.Asynchronousupdating

  23. Hebbian Learning Probe pattern Dynamical evolution

  24. A Simple Example Step_1.Design a network with memorized patterns (vectors)[ 1, -1, 1 ] & [ -1, 1, -1 ]

  25. #1: y1 #2: y2 #3: y3 Step_2. Initialization There are 8 different states that can be reached by the net and therefore can be used as its initial state

  26. Step_3. Iterate till convergence 3 different examples of the net’s flow - Synchronous Updating - It converges immediately

  27. Step_3. Iterate till convergence Stored pattern - Synchronous Updating - Schematic diagram of all the dynamical trajectories that correspond to the designed net.

  28. Or Step_3. Iterate till convergence - Asynchronous Updating - Each time, select one neuron at random and update its state with the previous rule and the –usual- convention that if the total input to that neuron is 0 its state remains unchanged

  29. Explanation of the convergence There is an energy function related with each state of the Hopfield network E( [y1, y2, …, yn]T ) = -Σ Σ wij yiyj where [y1, y2, …, yn]T is the vector of neurons’ output, wijis the weight from neuron j toneuron i, and the double sum is over i and j.

  30. The corresponding dynamical system evolves toward states of lower Energy

  31. States of lowest energy correspond to attractors of Hopfield-net dynamics Attractor-state E( [y1, y2, …, yn]T)= = -Σ Σ wij yiyj

  32. Capacity of the Hopfield memory In short, while training the net (via the outer-product rule) we’re storing patterns by posing different attractors in the state-space of the system. While operating, the net searches the closest attractor. When this is found, the corresponding pattern of activation is outputted

  33. How many patterns we can store in a Hopfield-net ? 0.15 N,N: # neurons

  34. Computer Experimentation Class-project A simple Pattern Recognition Example

  35. Stored Patterns (binary images)

  36. Perfect Recall- Image Restoration Erroneous Recall

  37. Irrelevant results Note: explain the ‘negatives’ ….

  38. The continuous Hopfield-Net as optimization machinery

  39. [ Tank and Hopfield ; IEEE Trans. Circuits Syst. 1986; 33: 533-541.]: ‘Simple "Neural" Optimization Networks: An A/D Converter, Signal Decision Circuit, and a Linear Programming Circuit’

  40. Hopfield modified his network so as to work with continuous activation and -by adopting a dynamical-systems approach- showed that the resulting system is characterizedby a Lyaponov-function who termed it ‘Computational-Energy’& which can be used to tailor the net for specific optimizations

  41. The system of coupled differential equation describing the operation of continuous Hopfield net Neuronal outputs: Yi≡ Vi Biases: Ii Weights: Wij ≡ Tij Tij=Tji και Tij=0  The Computational Energy

  42. When Hopfield nets are used for functionoptimization, the objective function Fto be minimized is written as energy functionin the form of computational energy E . The comparison betweenE and Fleadsto the design, i.e. definition of links and biases, of the networkthat can solve the problem.

  43. The actual advantage of doing this is that the Hopfield-net has a direct hardware implementation that enables even a VLSI-integration of the algorithm performing the optimization task

  44. An example: ‘Dominant-Mode Clustering’ Given a set of N vectors {Xi} define the k among them that form the most compact cluster {Zi} The objective function F can be written easily in the form of computational energy E

  45. There’s an additional Constraint so as kneurons are ‘on’ With each pattern Xi we associate a neuron in the Hopfield network ( i.e. #neurons = N ). The synaptic weights are the pairwise-distances (*2) If its activation is ‘1’ when the net will converge the corresponding pattern will be included in the cluster.

  46. A classical example: ‘The Travelling Salesman Problem’

More Related