message8

Date: Fri, 30 May 2003 07:27:42 -0500 (CDT)

From: Bill Hibbard <test@demedici.ssec.wisc.edu>

Reply-To: sl4@sl4.org

To: sl4@sl4.org

Subject: Re: SIAI's flawed friendliness analysis

On Thu, 29 May 2003, Eliezer S. Yudkowsky wrote:

> "Happiness in human facial expressions, voices and body language, as

> trained by human behavior experts".

>

> Not only does this one get satisfied by euphoride, it gets satisfied by

> quintillions of tiny little micromachined mannequins. Of course, it will

> appear to work for as long as the AI does not have the physical ability to

> replace humans with tiny little mannequins, or for as long as the AI

> calculates it cannot win such a battle once begun. A nice, invisible,

> silent kill.

The essence of intelligence is a simulation model of the world

that is used to predict the long-term effect of behavior on

values, and hence solve the credit assignment problem for

reinforcement learning. A tight stimulus-response loop

satisfying immediate values has nothing to do with intelligence.

An intelligent mind will develop a model of the world that

equates human happiness with a loving family life, adequate

food and shelter, physical exercise, freedom, a meaningful

vocation, friends, etc. And it will equate human unhappiness

with abusive relations, loneliness, homelessness, hunger, lack

of freedon, poor health, drug addiction, etc. Its behavior

will be based on this model, trying to promote the long-term

happiness of humans.

Human babies love their mothers based on simple values about

touch, warmth, milk, smiles and sounds. But as the baby's

mind learns, those simple values get connected to a rich set

of values about the mother, via a simulation model of the

mother and surroundings. This elaboration of simple values

will happen in any truly intelligent AI. I can't speak for

others, but what rocks my boat is the sound of happiness in

my wife's voice - a simple, grounded value. Not for a second

does my model of the world suggest giving her drugs or

replacing her with a doll and a tape recording.

Your argument can be applied against any grounded defintion

of values. Hence the SIAI guidelines define a supergoal that

is not grounded, and hence allows whatever interpretation an

AI designer wants to apply. Similarly, various other SIAI

recommendations use lots of value words without defining them.

Lack of grounding for value words in safe AI guidelines are

intended to protect against the imagined dangers of a

non-intelligent stimulus-response loop satisfying immediate

values, but have the actual effect of providing a loophole to

those with motives to circumvent safe AI guidelines.

> There is no divine right of democracy; it does not confer infallibility.

> What it does confer is faith and the illusion of infallibility. Congress

> is not capable of understanding how little it knows, which is what makes

> it dangerous. Democracy has known bugs; and those bugs, applied to

> Singularity scenarios, result in predictable kills - one of which you

> have just demonstrated.

Thanks for being honest about this. Everyone will have to decide

for themselves whether they trust their elected representatives

or the SIAI.

> Incidentally, Hibbard, if I were given to trying to pass regulations,

> which I'm not, I'd prohibit the use of your reinforcement architecture,

> which is, of course, invariably fatal...

Prohibiting reinforcement learning would ensure safety by

eliminating intelligence.

----------------------------------------------------------

Bill Hibbard, SSEC, 1225 W. Dayton St., Madison, WI 53706

test@demedici.ssec.wisc.edu 608-263-4427 fax: 608-263-6738

http://www.ssec.wisc.edu/~billh/vis.html