RE: Yudkowsky's AI (again)

Lyle Burkhead (lybrhed@earthlink.net)
Tue, 30 Mar 1999 23:04:26 -0800

Eliezer writes,

> To the extent that I can define Singularitarianism,
> it *is* about making money. It takes time, brains, and money
> to get to the Singularity, and we have less money
> than time and brains, so presently being a Singularitarian
> is about money. Take that how you like.

Well, it doesn't bother me. Maybe we disagree less than I thought. I don't use the word Singularity, but my ultimate aim is probably similar to yours. Raising money isn't a problem for me at this stage of my life; I am pursuing my own projects with my own money. But I certainly agree that action is essential, and action requires money.

I wonder: why do you think it is necessary to pursue your project under the aegis of Extropianism? -- or Singularitarianism, which sounds even more cultish? There is another meme that *is* about making money, and that meme plays a more fundamental role in your life than Extropianism. Maybe a simple reframe is all you need. Your project could be conceived in a different context, described in different terms, and carried out under a different aegis. As I said in one of my last posts before dropping off the list two years ago, Extropianism is just a watered down version of the ur-meme. Why not go back to the real thing? Your own people will give you all the money you need, if you explain what you want to do in terms that make sense to them. That would be one way to set events in motion.

Lyle