From: GBurch1@aol.com
Date: Fri Sep 18 1998 - 17:10:28 MDT
In a message dated 98-09-18 18:37:41 EDT, Robin Hanson wrote:
> Does *anyone* reading this other than
> Eliezer think that Eliezer's first post constitutes such a persuasive
> technical analysis?
Isn't the problem that a key term in all of these discussions, "intelligence"
(appearing in such other key terms as "human equivalence", "augmentation",
etc.), is simply not well understood at this point? The proponents of a
"singularity" all seem to have this concept built into their basic schema of
super-acceleration in the sense that they assume that "intelligence" can be
replicated in a machine and then, once replicated, enhanced. I'm sure that
essentially every subscriber to this list would agree with that general
assertion. But, until this essential factor is more subject to
quantification, I doubt that analysis of the rigor you are seeking, Robin, is
possible. That doesn't make the discussion pointless, it just makes it less
precise than we might like.
Greg Burch <GBurch1@aol.com>----<burchg@liddellsapp.com>
Attorney ::: Director, Extropy Institute ::: Wilderness Guide
http://users.aol.com/gburch1 -or- http://members.aol.com/gburch1
"Good ideas are not adopted automatically. They must
be driven into practice with courageous impatience."
-- Admiral Hyman Rickover
This archive was generated by hypermail 2.1.5 : Fri Nov 01 2002 - 14:49:35 MST