Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Here's a few basic reasons:

a) There are business incentives to give the AI more capabilities and more information about the world.

For example, if you have a cheeseburger bun factory with a complicated production line then you could use the AI to come up with more efficient layouts and machines for making cheeseburger buns. After awhile you realize that humans implementing what the AI explained is by far the slowest part of the pipeline, and ask if there's some way to have the AI build the production line by having it directly do the online ordering or 3d printing or whatever. At first you pay people to sanity-check what the AI is ordering and doing, but they never notice any problems so eventually you cut costs by laying them off.

b) Side channels and exploits are a thing.

Imagine yourself in the AI's place [1]. Is there really nothing you can do? Or is the security based around the fact that you don't try to escape?

"Hey Larry, why is the description for so many orders coming out as 'Warning: ketchup parameter set too high. See <a href="javascript:{...sudo...WebSocket...dumps.wikimedia.org.../resetpassword/?user=billgates...}">the manual</a>.'?".

"God damnit Ted, do you not see the link that says 'the manual'? Maybe you should click that before bothering me!"

c) Instrumental goals.

Any optimization process with an unqualified goal like "find a way to use this API to make lots of burgers, then do that" will favor plans where said process ends up with huge amounts of power over the world; simply because more burgers get made in that situation than in other situations. In some senses, failing to actively search for escape exploits could be seen as a design flaw because the algorithm is failing to find clever solutions that are better according to the stated goal.

1: http://lesswrong.com/lw/qk/that_alien_message/



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: