RE: Yudkowsky's AI (again)

From: Lyle Burkhead (lybrhed@earthlink.net)
Date: Wed Mar 31 1999 - 00:04:26 MST


Eliezer writes,

> To the extent that I can define Singularitarianism,
> it *is* about making money. It takes time, brains, and money
> to get to the Singularity, and we have less money
> than time and brains, so presently being a Singularitarian
> is about money. Take that how you like.

Well, it doesn't bother me. Maybe we disagree less than I thought. I
don't use the word Singularity, but my ultimate aim is probably similar
to yours. Raising money isn't a problem for me at this stage of my
life; I am pursuing my own projects with my own money. But I certainly
agree that action is essential, and action requires money.

I wonder: why do you think it is necessary to pursue your project under
the aegis of Extropianism? -- or Singularitarianism, which sounds even
more cultish? There is another meme that *is* about making money, and
that meme plays a more fundamental role in your life than Extropianism.
Maybe a simple reframe is all you need. Your project could be conceived
in a different context, described in different terms, and carried out
under a different aegis. As I said in one of my last posts before
dropping off the list two years ago, Extropianism is just a watered down
version of the ur-meme. Why not go back to the real thing? Your own
people will give you all the money you need, if you explain what you
want to do in terms that make sense to them. That would be one way to
set events in motion.

Lyle



This archive was generated by hypermail 2.1.5 : Fri Nov 01 2002 - 15:03:26 MST