Re: Singularity: Just Say No!

From: Eliezer S. Yudkowsky (sentience@pobox.com)
Date: Tue Jan 26 1999 - 21:56:17 MST


"J. R. Molloy" wrote:
>
> A sane seed AI presents more of a threat to humanity than does an insane AI
> because a sane AI would function at extreme variance to the insane human
> cultures which prevail on Earth. No joke.
>
> Cheers,
>
> J. R.

OK, basically I have three comments here.

One, if it's really that bad, I'm going to side with the AIs.
It's the sane thing to do.

Two, an insane AI would be insane in a totally different way than _us_,
and might function at a variance that would not only be extreme, but
also randomly destructive.

Three, no matter how insane we are, we're still trying. I think there
really are a few things we've got right. We may still be insane, but
we're saner than in medieval times. A sane AI would have at least that
much in common. An insane AI would have nothing in common at all.

-- 
        sentience@pobox.com         Eliezer S. Yudkowsky
         http://pobox.com/~sentience/AI_design.temp.html
          http://pobox.com/~sentience/sing_analysis.html
Disclaimer:  Unless otherwise specified, I'm not telling you
everything I think I know.


This archive was generated by hypermail 2.1.5 : Fri Nov 01 2002 - 15:02:56 MST