1 / 16

AI Safety

AI Safety. CMPT 310 Summer 2019. AGIs. A rtificial G eneral I ntelligence. Superintelligence. [a superintelligence is] any intellect that greatly exceeds the cognitive performance of humans in virtually all domains of interest -- Nick Bostrom. Superintelligence.

nrodriguez
Download Presentation

AI Safety

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. AI Safety CMPT 310 Summer 2019

  2. AGIs Artificial General Intelligence

  3. Superintelligence [a superintelligence is] any intellect that greatly exceeds the cognitive performance of humans in virtually all domains of interest -- Nick Bostrom

  4. Superintelligence one example of superintelligence would be a human mind running on upgraded hardware, e.g. say 3 million times faster than the average human brain

  5. Superintelligence whether or not superintelligences are conscious, or truly intelligent, doesn’t matter --- what we care about is what they end up doing

  6. The Last Invention … the first ultraintelligent machine is the lastinvention that man need ever make, provided that the machine is docile enough to tell us how to keep it under control. -- I.J. Good

  7. Time to Worry? Should we be worried about creating AGIs, or superintelligences, that might not work they way we want?

  8. Example: Minimizing Human Suffering • Suppose we create an AGI with the goal to minimize human suffering • Suppose it concludes that, even in paradise, human nature is such that humans will always find some way to make themselves suffer • Therefore, the agent concludes that the best way to minimize human suffering is to … kill all humans

  9. Example: Paper Clip Maximizer Suppose we have an AI whose only goal is to make as many paper clips as possible. The AI will realize quickly that it would be much better if there were no humans because humans might decide to switch it off. Because if humans do so, there would be fewer paper clips. Also, human bodies contain a lot of atoms that could be made into paper clips. The future that the AI would be trying to gear towards would be one in which there were a lot of paper clips but no humans. -- Nick Bostrom

  10. Time to Worry? … AI does not hate you, nor does it love you, but you are made out of atoms which it can use for something else. -- Eliezer Yudkowsky

  11. Asimov’s 3 Laws of Robotics First Law A robot may not injure a human being or, through inaction, allow a human being to come to harm. Second Law A robot must obey the orders given it by human beings except where such orders would conflict with the First Law. Third Law A robot must protect its own existence as long as such protection does not conflict with the First or Second Laws.

  12. A Common Theme AGIs might kill us!

  13. AI Safety Some people suggest that we need to be thinking about safety issues when designing AI systems. Other people say it’s like worrying about over-population on Mars --- we are so far away from AGI that there’s no need to worry about it yet.

  14. Corrigibility one suggestion is that we ought to design systems that are corrigible, i.e. systems that know they are not always right and so shouldn’t just blindly do what they think is best; they must be open to correction

  15. Value Alignment another suggestion is that we somehow must ensure that AGIs share our most important human values

  16. Stop Buttons In industry, dangerous machines have a stop button that a person can press to immediately turn off the machine. Perhaps we should be adding stop buttons to our AI systems? Check out this video on the stop button problem …

More Related