Re: Why would AI want to be friendly?

From: Franklin Wayne Poley (culturex@vcn.bc.ca)
Date: Mon Sep 25 2000 - 16:28:49 MDT


On Mon, 25 Sep 2000, Eliezer S. Yudkowsky wrote:

> Franklin Wayne Poley wrote:
> >
> > I have given hundreds of IQ tests over the course of my career and
> > participated in the development of one of them (Cattell's CAB). If I were
> > to measure transhuman-machine intelligence and human intelligence; and
> > compare the profiles, how would they differ?
>
> The transhuman would max out every single IQ test. It is just barely possible
> that a mildly transhuman AI running on sufficiently limited hardware might
> perform badly on a test of visual intelligence, or - if isolated from the
> Internet - of cultural knowledge. A true superintelligence would max those
> out as well.
>
> The transhuman can beat the living daylights out of you at chess or Go or
> poker, and do the same to Deep Blue and Kasparov with scarcely more effort.
> Ve can hack source code, prove the Riemann Hypothesis, win a debate, offer
> psychiatric counseling, author a scientific paper, design experimental
> procedures, write a poem, paint a picture, and create new technologies. Any
> other questions?
>
> -- -- -- -- --
> Eliezer S. Yudkowsky http://singinst.org/
> Research Fellow, Singularity Institute for Artificial Intelligence

Yes. According to "Futurama" which is where I get my better ideas on
machine psychology, John Quincy Adding Machine was the first
Robot-for-President. According to your web site, "final-stage AI" will
reach "transhumanity...probably around 2008 or 2010". In which election
will you first run John Quincy Adding Machine?
FWP



This archive was generated by hypermail 2b29 : Mon Oct 02 2000 - 17:39:01 MDT