1 / 19

COGNITIVE MEMORY HUMAN AND MACHINE

COGNITIVE MEMORY HUMAN AND MACHINE. by BERNARD WIDROW JUAN CARLOS ARAGON INFORMATION SYSTEMS LABORATORY DEPT. OF ELECTRICAL ENGINEERING STANFORD UNIVERSITY June, 2008. THE 3 HOWS.

stacye
Download Presentation

COGNITIVE MEMORY HUMAN AND MACHINE

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. COGNITIVE MEMORYHUMAN AND MACHINE by BERNARD WIDROW JUAN CARLOS ARAGON INFORMATION SYSTEMS LABORATORY DEPT. OF ELECTRICAL ENGINEERING STANFORD UNIVERSITY June, 2008

  2. THE 3 HOWS How does human memory work ? How could I build a memory like that ? How could I use it to solve practical problems ?

  3. What would we like to do ? • Design a memory system that is as simple as possible, but behaves like and emulateshuman memory. Why would we like to do this ? • To develop a new kind of memory for computers, adjunct to existing forms of memory, to facilitate solutions to problems in artificial intelligence, pattern recognition, speech recognition, control systems etc. • To advance cognitive science with new insight into the working of human memory.

  4. SALIENT FEATURES OF COGNITIVE MEMORY • Stores sensory patterns (visual, auditory, tactile; radar, sonar, etc.). • Stores patterns wherever space is available, not in specified memory locations. • Stores simultaneously sensed input patterns in the same folder (e.g., simultaneous visual and auditory patterns are stored together). • Data recovery is in response to “prompt” input patterns (e.g., a visual or auditory input pattern would trigger recall). • Autoassociative neural networks are used in the data retrieval system.

  5. Satellite photo of Diego Garcia Island showing U.S. Air Force base

  6. Aircraft parked in a area near the main runway

  7. Scanning, looking for a hit

  8. This is a hit, object is KC135

  9. MEMORY OUTPUT PATTERNS PATTERN STORAGE ( TRAINING ) MUX PATTERN RETRIEVAL ( SENSING ) NN MUX HIT? NN V C VC = VISUAL CORTEX CAMERA HIT? V C BUFFER CAMERA A SIMPLE COGNITIVE MEMORY FOR PATTERN RECOGNITION

  10. THREE PHOTOS OF BERNARD WIDROW USED FOR TRAINING A PHOTO OF JUAN CARLOS ARAGON, VICTOR ELIASHBERG, AND BERNARD WIDROW USED FOR SENSING

  11. FACE DETECTION • Training (low resolution, 20x20 pixel images) • One image of a person’s face was trained in • The image was adjusted by • Rotation (2° increments, 7 angles) • Translation (left/right, up/down, 1 pixel increments, 9 positions) • Brightness (5 levels of intensity) • Total number of training patterns = 315 • Training time 12 minutes on AMD 64 bit Athlon 2.6 GHz computer for 0.25% MSE

  12. FACE DETECTION • Sensing (low resolution, 20x20 pixel images) • Each input pattern was adjusted by • Scaling (6 window sizes) • Translation (90 pixel increments) • Errors with background were ~8X greater than with a person’s face • 60 patterns per second through neural network • Autoassociative neural network has total of 1100 neurons distributed over 3 layers • 400 neurons, 400 weights per neuron, first layer • 300 neurons, 400 weights per neuron, second layer • 400 neurons, 300 weights per neuron, third layer

  13. FACE RECOGNITION • Training (high resolution, 50x50 pixel images) • All 3 images of Widrow’s face were trained in • Each image was adjusted by • Rotation (2° increments, 7 angles) • Translation (left/right, up/down, 1 pixel increments, 25 positions) • Scaling (3 window sizes) • Total number of training patterns = 1575 • Training time 2.6 hours in AMD 64 bit Athlon 2.6 GHz computer for 0.25% MSE

  14. FACE RECOGNITION • Sensing (high resolution, 50x50 pixel images) • Each input pattern was adjusted by • Scaling (6 window sizes) • Translation (2 pixel increments, 25 positions) • Brightness (6 levels of intensity) • Optimization was done for each detected face • Errors with unidentified faces were ~4X greater than with Widrow’s face • 5 patterns per second through neural network • Autoassociative neural network • 1800 neurons, 2500 weights per neuron, first layer • 1500 neurons, 1800 weights per neuron, second layer • 2500 neurons, 1500 weights per neuron, third layer • Total 5800 neurons, 10,950,000 weights

  15. (a) (b) SENSING PATTERNS OBTAINED FROM WIDROW’S FACE WITH TWO WINDOW SIZES, (a) STRAIGHT UP, AND (b) ROTATED

  16. Cognitive Memory Challenged • Photographs distributed by NIST for the Face Recognition Grand Challenge version 1 were used for training and testing. • Photographs of 75 persons were selected for training. • 75 photographs NOT trained-in of the above persons were selected for sensing purposes. • 300 photographs of persons NOT trained-in were selected for sensing. • In total, 75 photographs were used for training and 375 for sensing. • Autoassociative neural network had 3 layers distributed as follows: • 2000 neurons in the first layer • 1500 neurons in the second layer • 10000 neurons in the last layer • Total number of weights: 38 million. Retina size: 100 × 100 pixels. • Results: 75 people trained-in were recognized and identified without error, while the 300 people not trained-in were rejected by the Cognitive Memory system.

  17. VIDEO ON AIRCRAFT IDENTIFICATION

  18. FACE RECOGNITION VIDEO

More Related