An alternate stance on moderation (from @Habryka.)
This is from this comment responding to this post about there being too many bans on LessWrong. Note how the LessWrong is less moderated than here in that it (I guess) responds to individual posts less often, but more moderated in that I guess it rate limits people more without reason.
I found it thought provoking. I'd recommend reading it.
Thanks for making this post!
One of the reasons why I like rate-limits instead of bans is that it allows people to complain about the rate-limiting and to participate in discussion on their own posts (so seeing a harsh rate-limit of something like "1 comment per 3 days" is not equivalent to a general ban from LessWrong, but should be more interpreted as "please comment primarily on your own posts", though of course it shares many important properties of a ban).
This is a pretty opposite approach to the EA forum which favours bans.
...Things that seem most important to bring up in terms of moderation philosophy:
Moderation on LessWrong does not depend on effort
"Another thing I've noticed is that almost all the users are trying. They are trying to use rationality, trying to understand what's been written here, trying to apply Baye's rule or understand AI. Even some of the users with negative karma are trying, just having more difficulty."
Just because someone is genuinely trying to contribute to LessWrong, does not mean LessWrong is a good place for them. LessWrong has a particular culture, with particular standards and particular interests, and I think many people, even if they are genuinely trying, don't fit well within that culture and those standards.
In making rate-limiting decisions like this I don't pay much attention to whether the user in question is "genuinely
This post was partly inspired by, and shares some themes with, this Joe Carlsmith post. My post (unsurprisingly) expresses fewer concepts with less clarity and resonance, but is hopefully of some value regardless.
Content warning: description of animal death.
I live in a ...
Hey Sam — thanks for this really helpful comment. I think I will do this & do so at any future places I live with wool carpets.
I am thrilled to introduce EA in Arabic (الإحسان الفعال), a pioneering initiative aimed at bringing the principles of effective altruism to Arabic-speaking communities worldwide.
Spoken by more than 400 million people worldwide, Arabic plays a pivotal role...
Thank you Elham, I'm so happy to see your comment!
Honestly, this is an issue that I'm kinda struggling with too. Would you like to have a quick call to discuss our experiences and maybe collaborate on something?
People around me are very interested in AI taking over the world, so a big question is under what circumstances a system might be able to do that—what kind of capabilities could elevate an entity above the melange of inter-agent conflict and into solipsistic hegemony?
We...
Executive summary: AI systems with unusual values may be able to substantially influence the future without needing to take over the world, by gradually shifting human values through persuasion and cultural influence.
Key points:
Self-evaluation using LLMs is used in reward modeling, model-based benchmarks like GPTScore and AlpacaEval, self-refinement, and constitutional AI. LLMs have been shown to be accurate at approximating human annotators on some tasks.
But these methods are threatened by self...
Executive summary: Frontier language models exhibit self-preference when evaluating text outputs, favoring their own generations over those from other models or humans, and this bias appears to be causally linked to their ability to recognize their own outputs.
Key points:
Excerpt from Impact Report 2022-2023
"Creating a better future, for animals and humans
£242,510 of research funded
3 Home Office meetings attended
5 PhDs completed through the FRAME Lab
33 people attended our Training School in Norway and our experimental design training...
Executive summary: FRAME (Fund for the Replacement of Animals in Medical Experiments) is an impactful animal welfare charity working to end the use of animals in biomedical research and testing by funding research into non-animal methods, educating scientists, and advocating for policy changes.
Key points:
Anders Sandberg has written a “final report” released simultaneously with the announcement of FHI’s closure. The abstract and an excerpt follow.
...Normally manifestos are written first, and then hopefully stimulate actors to implement their vision. This document is the reverse
Executive summary: The Future of Humanity Institute (FHI) achieved notable successes in its mission from 2005-2024 through long-term research perspectives, interdisciplinary work, and adaptable operations, though challenges included university politics, communication gaps, and scaling issues.
Key points:
You said "Ruling out Z first seems more plausible, as Z negatively affects the present people, even quite strongly so compared to A and A+." The same argument would support 1 over 2.
Granted, but this example presents just a binary choice, with none of the added complexity of choosing between three options, so we can't infer much from it.
...Then you said "Ruling out A+ is only motivated by an arbitrary-seeming decision to compare just A+ and Z first, merely because they have the same population size (...so what?)." Similarly, I could say "Picking 2 is onl
I started working in cooperative AI almost a year ago, and as an emerging field I found it quite confusing at times since there is very little introductory material aimed at beginners. My hope with this post is that by summing up my own confusions and how I understand them...
Thank you Shaun!
I found myself wondering where we would fit AI Law / AI Policy into that model.
I would think policy work might be spread out over the landscape? As an example, if we think of policy work aiming to establishing the use of certain evaluations of systems, such evaluations could target different kinds of risk/qualities that would map to different parts of the diagram?
It seems plausible to me that those involved in Nonlinear have received more social sanction than those involved in FTX, even though the latter was obviously more harmful to this community and the world.