Brian_Tomasik comments on S-risk FAQ - Effective Altruism Forum

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (9)

You are viewing a single comment's thread. Show more comments above.

Comment author: Brian_Tomasik 19 September 2017 12:11:11AM 3 points [-]

the sort of thing we were pointing at in the late 90s before we started talking about x-risk

I'd be interested to hear more about that if you want to take the time.

Comment author: gworley3  (EA Profile) 19 September 2017 12:50:18AM 4 points [-]

My memory is somewhat fuzzy here because it was almost 20 years ago, but I seem to recall discussions on Extropians about far future "bad" outcomes. In those early days much of the discussion focused around salient outcomes like "robots wipe out humans" that we picked up from fiction or outcomes that let people grind their particular axes (capitalist dystopian future! ecoterrorist dystopian future! _ dystopian future!), but there was definitely more serious focus around some particular issues.

I remember we worried a lot about grey goo, AIs, extraterrestrial aliens, pandemics, nuclear weapons, etc. A lot of it was focused on getting wiped out (existential threats), but some of it was about undesirable outcomes we wouldn't want to live in. Some of this was about s-risks I'm sure, but I feel like a lot of it was really more about worries over value drift.

I'm not sure there's much else there, though. We knew bad outcomes were possible, but we were mostly optimistic and hadn't developed anything like the risk-avoidance mindset that's become relatively more prevalent today.

Comment author: [deleted] 19 September 2017 02:07:40AM *  3 points [-]

Here is s-risk discussion from 2003: http://sl4.org/wiki/HyperExistentialRisk

You might be able to find more discussion by searching for "suffering" or other relative terms using <site:sl4.org>. You can also search the extropians archive <site:diyhpl.us/~bryan/irc/extropians/>, but it doesn't seem to be fully indexed.