Hide table of contents

Introduction

Ben Kuhn recently wrote on how and why effective altruism may or may not seem welcoming to newcomers, and suggestions as to what could be done to make effective altruism more welcoming. He noticed a lack of diversity (within effective altruism) on lines such as age, career, lifestyle, socioeconomic class, level and type of education, etc. He notes everyone else might feel crowded out by some perceived mono-culture, and while like any community effective altruism might tend toward this failure mode by default, it's in everyone's interest to go beyond typical thinking in brainstorming new ways to welcome others.

On some object-level issues Mr. Kuhn mentioned, debate broke out among commentators. Despite trying to quell or improve the quality of this debate as a moderator, Mr. Kuhn feels it degraded into acrimony, vitriol, and strawmanning. Indeed, some commentators expressed they might leave effective altruism behind entirely. He felt this fiasco discouraged him more than any explicit counter-argument that trying to build and hold together effective altruism as a singular community isn't worth it if there will be so much collateral damage to the movement and its adherents. Several dozen people appeared to agree with Mr. Kuhn's newfound pessimism, or at least they didn't disagree with it. I, on the other hand, strongly disagree. While I acknowledge effective altruism needs to look at itself, and thus strong optimism about its self-awareness may not be justified, I don't believe a doom-and-gloom outlook is justified either.

Below is my open letter to Ben Kuhn and effective altruism as a whole to not give up on the value of building community. It's been edited for brevity, and clarity. Note that I never originally discussed the object-level issues in debates which originally so discouraged Mr. Kuhn. I encourage you not to make your discussion of this open letter about that, as the pattern of falling into a debates in which we misrepresent each other is what urged this letter in the first place.


Hello Ben. I hope you're doing well. I'm Evan Gaensbauer.

I read your last blog post, and all the comments on it. It appears no more than five anonymous persons participated in these fights online you cited. Posting anonymously allows people to feel they can write comments online they otherwise wouldn't feel comfortable with.They may be expressing more extreme views than what they really hold in their everyday thoughts. Anyway, five anonymous people having a debate about one or two issues doesn't strike me as sufficient to stop trying to make the effective altruism community more welcoming. Maybe no one individual should try so hard to hold people together if it strains their willpower and capacity to handle stress. However, I don't think that's sufficient for all of us to stop trying, though. I don't want to get rid of having one thing called "the EA community".

Effective altruism could become a movement of multiple communities, in which people of different political stripes or cluster-groups stick to their own. However, I'd be fine with that if everyone kept donating money to effective charities, discussed and debated what those were, while still largely agreeing with one another, and kept thinking about how they can lead lifestyles and careers which are more effectively altruistic. I think such a movement could be welcoming, as multiple communities, because when effective altruism stops being correlated with particular ideologies, it stops being an ideology itself, and people would think of it as common-sense. It stops being perceived as politics, and starts being perceived as the sensible mechanics of doing good. That would be great news.

Five people debating in the comments on one blog? I bet, collectively, they spent less than five person-hours writing those comments. I can scarcely think of one individual in this community who hasn't spent dozens, if not hundreds, of hours learning more about effective altruism, ethics, and research to inform not just where they donate, but how to inject more altruism into lots of parts of their lives. That effort trumps political debate. For whatever reason, five people were drawn to one blog time and again over the course of five days, likely there for only minutes at a time, to have a debate which was refereed, rather than just deleted. These people weren't calling each other names, or swearing, or launching ad hominem attacks all the time. That's what I think of as "calm" and "level-headed". If you're in a bubble of people like that, even the most pessimistic people drawn to flame wars are more polite to one another than people I've witnessed arguing my entire life.

We can't measure how much "acrimony" is captured in these comments. Maybe these people went on with their days, slightly grumpy, but in a regular mood within the hour. Maybe not. We don't know, though, and I don't think that's substantial enough evidence to become so pessimistic. Surely someone was effected, as one or two commentators wrote they might leave effective altruism if it continues to be like this. If having these conversations in person isn't an option because these anonymous commentators aren't in the Bay Area, you or I could reach out personally, let them know an email or something, and chat with them on Skype. That takes the misinterpretation and misplaced vitriol out of Internet debates. Based on how dearly they consider their relationship with effective altruism, I imagine they're people we could trust personally and in discussion, and people we might already know.

If people feel scared, threatened, and vulnerable, those are things which make effective altruism unwelcoming, but can be addressed.

I have made hundreds of friends on Facebook through effective altruism. "Facebook friends" often means little, but what I mean is there are hundreds of individuals, strangers and acquaintances and role models, who want me to be part of their professional and personal networks, because they assume through association effective altruism makes us kindred spirits. I've met dozens of them in person. We lean on each other for emotional and moral support. Nobody is taking for granted the contributions of others. I'm honored people like yourself I aspire in my methods of thought and care to become more like treat me no less well than anyone else.

While there is a share of effective altruists who feel unwelcome, I think people can and do make enough friends and allies in this movement to keep their ear to the ground about effective altruism news. Never have I ever encountered someone who didn't feel welcomed by effective altruism, but then also abandoned their drive to do good works, or a desire for clearer thinking in assessing their own choices. Even if someone stops publicly affiliating with "effective altruism" the movement, if they still donate to effective charities, if they still make ethical lifestyle choices like vegetarianism or something else as it suits their values, then I think effective altruism has done it's job. In engaging the public, effective altruism's mission is use its passion to inspire confidence in others to take substantially effective action, not just symbolically signal association with a special group. Even if someone exits effective altruism, they can still donate to effective charities, they can still spread messages professionally and privately about quantifying efforts and evidence-based interventions, and they can still advocate. A piece of effective altruism can be left with them in all the good works they do for the rest of their lives.

I'm not saying we should be totally optimistic about making effective altruism more welcoming. Effective altruism will never be perfect, and more struggles may come. I just mean five commentators this one week among the thousands of people becoming involved with effective altruism doesn't seem enough evidence to justify massive pessimism either. Sincerely,

Evan Gaensbauer


Conclusion

When I wrote:

I'm honored people like yourself I aspire in my methods of thought and care to become more like treat me no less well than anyone else.

this extends to anyone now reading this as much as it does Ben Kuhn. I've met some people I once thought of as personal heroes, such as Will MacAskill, founder of Giving What We Can, and Holden Karnofsky, founder of Givewell. When I met them, they told me how my thoughts, and the rationale behind my actions, even if they wouldn't agree with them perfectly, they definitely respected and were grateful for. That's why I think of these people as role models I can aspire to be like, not great men on a pedestal who have a status I can never achieve. I reflected on how personal friends I knew before getting involved with effective altruism, like Joey Savoie and Xio Kikauka, who founded Charity Science, or others who have donated thousands of dollars, like Andrew McKnight, of volunteered dozens of hours, like Eric Chisholm and Max Carpendale, should be my role models, too.

Even very different people who I sense sincerely wish to make an effort, even if they can't always make the most effort, are kindred spirits in challenging themselves to do good even when they must make great effort to change their minds and actions to do so. It's this willingness which can act as a common core of effective altruism, feelings of solidarity and humility as being a peer alongside your own role models, which can also be the spirit with which we welcome others. We're not close to failing in that regard to give up hope, so let's keep trying.

5

0
0

Reactions

0
0

More posts like this

Comments18
Sorted by Click to highlight new comments since: Today at 1:30 PM

So like you, I’ve been inspired by lots of people in the EA movement and I’m optimistic that the EA movement can continue growing and do lots of good.

However, the history of disasters teaches us that major disasters are often preceded by minor disasters of a similar sort. If people took minor disasters more seriously and put stronger safeguards in place, there would be fewer major disasters. “Only the paranoid survive”, as Andy Grove says.

I’ve been really impressed by the reflectiveness and thoughtfulness of everyone I’ve met through the EA movement. But if the EA movement continues executing on its aggressive expansion plans, we’ll eventually start seeing a larger population of EAs that are less reflective by disposition than our current set. Someday they might even make up the majority of EAs.

Because of this, I think it’s worthwhile taking some time to figure out how to preserve our current culture (which I think is great for the most part) and guard against the failure modes that may come with growth. As Rob Bensinger writes, “The value of growing the EA movement is instrumental, and conditional.” If we’re ever going to codify our cultural norms, that feels like something we should do sooner rather than later. As the movement grows, it’s harder to get group buy-in on new proposed norms.

The biggest failure modes I foresee right now fall under the broad category of “politicization”. “Politicization” doesn’t refer to the discussion of traditionally political topics per se; rather, it refers to the dysfunctional conversational modes traditionally associated with political topics. In principle any topic can be politicized (e.g. the color of the sky in this short story).

Unfortunately, politicization is hard to talk about because discussing it tends to trigger object-level discussion of politicized topics. I’ve been thinking about it a fair amount lately, but I haven’t posted on it because I was afraid to trigger the wrong sort of discussion. However, now that you’ve brought the topic up, I may as well share some thoughts and hopefully contribute to a relatively nonpoliticized tone for this comment thread :)

So my first question is what the best way to ratchet politicization down is once it has been ratcheted up. It’s possible that, like cockroaches or government corruption, it’s a problem that’s difficult to get rid of once you acquire it. It’s also possible that expert mediators are relatively good at defusing politicization, and all the EA community needs is a team of respected neutral mediators (“flame war firefighters”) to deal with problems as they arise.

If there aren’t reliable ways to ratchet politicization down, then we’ll want to focus on prevention.

You write: “Effective altruism could become a movement of multiple communities, in which people of different political stripes or cluster-groups stick to their own.” I agree this would not necessarily be terrible, but it does seem potentially terrible. I can imagine a community of republican EAs and a community of democrat EAs, each deciding that the best way to be an effective altruist is to donate and promote “effectively altruistic” government policies (aka whichever policies they would have promoted anyway), and the two communities exactly cancel each other out with their work.

Right now we are not even close to the failure mode I describe above. The EA community currently has a strong norm against discussing government policy. I think this is a good norm and we shouldn’t let it degrade. There’s no question that improving policies can be high-impact, but if we allow discussion of policy on official EA channels, there’s nothing preventing that discussion from degrading in to a typical internet flamewar as the movement grows. If the EA movement ever wants to work on improving policy, we should do it very carefully.

Another way politicization could happen is if certain causes within the EA movement amp up their rhetoric and aggressively go around trying to convince EAs that their cause is the best.

This could trigger a corresponding response from other causes within the EA movement, and the official EA discussion channels could be flooded with cause prioritization discussion in the resulting arms race. I think we should be ready to ban cause prioritization discussion on mainstream EA channels and restrict it to dedicated cause prioritization channels if this starts happening. (E.g. a regular cause prioritization thread on this forum.)

Even worse, a heated cause prioritization discussion favors simple, legible accusations of the form “you’re a terrible person if you don’t support my cause” over the careful, reflective analysis that I currently see in the EA community. I see EA’s culture of careful, reflective analysis as the primary thing differentiating us from regular non-EA altruism. If our standards of discussion degrade, we may lose what makes us different.

I don’t think there are easy solutions but being aware of potential problems seems valuable.

The EA community currently has a strong norm against discussing government policy.

Does it? The rationality community has this norm because Politics Are The Mind-Killer, but I don't think it exists that much in EA. For instance, didn't the Global Priorities Project (or something) give policy advice to the UK government? And surely people have discussed GiveWell's criminal justice reform work.

Yeah, maybe it's not a "strong" norm, but I do perceive some norm and I think it's valuable. I think I'm much more OK with EA organizations like Givewell doing policy research and delivering targeted policy recommendations vs casual policy discussion in the EA Facebook group. Maybe because an individual organization is more likely to follow all the recommendations you describe here: employees discuss issues face-to-face in small groups, and are typically on sufficiently good terms to stay away from character assassinations.

BTW, I'm tempted to nominate you as "mediator of online EA discussions" to fight diffusion of responsibility in flamewar firefighting, but I guess your revealed preferences suggest that it's not a role that you particularly enjoy? It's not a role that I particularly enjoy either. I feel like part of the problem is that trying to carefully evaluate all sides of an issue and not be a jerk to everyone is a lot less fun than trying to win points for your side, so that's why the "honest middle" tends to leave a lot of politicized conversations and let extremists to duke it out. (You probably have the moral authority to nominate someone else if you wanted to though. Note that a "mediator" and a "moderator" are not quite the same thing... a "moderator" censors discussion that gets off track, whereas a "mediator" participates in discussions and tries to help each side understand the other. It feels like a bad idea for both roles to be played by the same person. I think ideally the EA community would have a sizable population of mediator types in order to keep us cohesive. We want disagreements to be about facts, not personalities.)

I'm flattered by your nomination!

I don't particularly enjoy flamewar firefighting, but it's not like I can't do it; it's just frustrating, distracting and not very immediately rewarding, and I don't feel very qualified even when I'm at the top of my game (let alone in the middle of a flamewar)! But you're not the first person to suggest this to me, so maybe I should update.

It seems pretty possible to ratchet down politicized discussion if you're careful about how you do it. Here are some general suggestions for fighting politicization:

  • Talk about things in person if possible. When talking online, I've noticed that I often feel offended/attacked by remarks where, if someone had changed like two words of the remark with something else, I would have been fine with it (e.g. "I think you're wrong to believe " vs. "it's disturbing that you believe "). For some reason this doesn't seem to happen nearly as often in person, probably because it's higher bandwidth.

  • Talk about things one-on-one or in small groups if possible. On polarized topics, I sometimes have the experience of saying something centrist, then being attacked by people who don't like side A because I forgot to put in an anti-side-A caveat, and then being attacked by people who don't like side B because I forgot to put in an anti-side-B caveat and the anti-side-A caveat made me look to much like a side-B-person to them.

    This can happen even if sides A and B are both essentially strawman positions that nobody in the discussion believes: the problem isn't that anyone is a tribalist for A or B, just that people are worried about running into A- or B-tribalists and argue with you to make sure you're not one. In a small enough group, it's much easier to remember which caveats you have to insert and which ones you can let people infer.

  • Be careful to separate people's arguments from people's character. As I recently wrote in the Facebook group (in response to someone saying "it's sad to see people talking about here" and "your attitude about Y is troubling"):

    One can also correct people's misconceptions without shaming them--e.g., by being careful to imply that they're incorrect simply by being misinformed rather than out of some kind of malice or character flaw (which indeed they often are, though not always). For example, instead of saying "it's sad that you believe "--which immediately puts someone on the defensive, because now if they admit that you're right then they also have to admit that your judgment of their character was correct--you can say "I don't think is true." Instead of saying " is troubling"--which people sometimes interpret as not only meaning that you think they're wrong but that you think they're a bad person for believing it!--you can say " suffers from " or " is actually harmful to express for ."

    I'm not arguing that you should believe the second statement over the first, only that purely as an empirical matter it seems to create a more effective and thoughtful discussion. And I'm also not necessarily arguing that one should never shame, just that if one is specifically worried about not-shaming people, it's possible to dissent without shaming.

These seem like good suggestions. I have my own list of suggestions here if anyone is interested. I think it would be valuable to institutionalize some set of suggestions by e.g. adding them to the set of links here. (I also like Paul Graham's disagreement hierarchy and your useful ideas in debate. And Scott Adams' idea of shorthand terms for debate tactics like "outragism" that we want to discourage.)

The EA community currently has a strong norm against discussing government policy.

does seem potentially terrible. I can imagine a community of republican EAs and a community of democrat EAs, each deciding that the best way to be an effective altruist is to donate and promote “effectively altruistic” government policies (aka whichever policies they would have promoted anyway), and the two communities exactly cancel each other out with their work.

I think we do not have anywhere near as strong a norm against this as is required. LW has it, though the line is constantly tested. In effective altruism, by contrast, we have Will MacAskill suggesting that supporting the Labour party might be the most effective thing to do (if you restricted yourself to domestic causes).

Can you remind me why we should rule out political issues from discussions about what is likely to be most effective please?

Sure. As LW has long held, politics is the mindkiller. Even the most rational people apply significantly lower epistemic standards to political issues, and generally end up endorsing pretty much the positions you'd expect them to on the basis of arational reasons. A good recent example if you're on facebook would be the thread where Tyler asked for recommendations for speakers who could talk on Systemic Change, which quickly devolved into everyone simply naming their pet favourite politicians/activists.

Edit: Actually, there are even better examples of where politics lead to EAs wasting a huge amount of effort in pointless arguments, but they're so political I'm afraid if I mention them here people will start arguing about them again!

A secondary reason is that it is far more diversive than anything else we do. For most EA causes, it is reasonably uncontroversial that they are at least not actively bad - there are few pro-african-poverty, pro-increasing-meat-consumption, or pro-extinction activists. But this cannot be said of many political issues, which are extremely diversive. Michelle has previously written about this if I recall correctly.

A secondary reason is that it is far more diversive than anything else we do ... Michelle has previously written about this if I recall correctly.

Found it: Michelle responding to one particularly diversive example. Rob's reply outlines the standard response to this argument, namely that politically causes might happen to be much more effective than apolitical ones, so we will just have to accept the associated costs.

Tibetan monks I think used to have a tradition of a physical circle of rope that you could step into to debate openly - the idea being that you leave offense and bad feelings about what's said in the ring in the ring after the debate so you can have a free and open discussion. This kind of norm might be helpful but difficult over the internet. The costs of controversy for political action, however, seem to me to be something you should just factor in to your decision.

Sure. So if that's the case, this discussion could be reframed as trying to estimate just how high those "controversy costs" are likely to be. My feeling is that the controversy costs are right-tailed, that is, there is a large probability of low/moderate costs but a small probability of large costs. Something that happens to controversies as they grow is that as they become more and more visible, it becomes harder and harder to avoid commenting on them, which reinforces their growth.

Personally, I think instead of working on policy in the near term, EA should think about how to think about policy... that is, why is it that policy discussions seem to predictably go wrong in a way that (say) effective malaria charity discussions don't, and how to fix that. If we find plausible solutions, that's valuable in order to

  1. clarify our own political thinking as a movement and make sure that whichever policies we push for are the right ones

  2. possibly improve the state of political discourse in general if our ideas are sufficiently compelling

  3. replace zero-sum political competition between ideologically opposed people with a sensible nonpoliticized policy discussion where peoples views become more accurate

In the EA Facebook thread Larks mentions, I said this and got 9 likes:

Aligning ourselves with almost any prominent political pundit at this stage risks alienating people who disagree with that pundit. EA is mostly liberal now; if we have a liberal speaker, that will make it even harder to interest conservatives. (Ideological diversity is the most valuable kind of diversity, so this matters.) Choosing a conservative or neutral pundit is a safer choice from this perspective. Tyler Cowen seems relatively sensible, non-polarizing, and thoughtful: video link

I’m optimistic that EA will be able to tackle political issues in the long run. But we have to do it right. In the same way none of us can be expected to choose charities as well as Givewell does in our spare time, none of us should assume that whatever political beliefs we’ve acquired in our spare time correlate all that well with the truth on political topics. (Note that politics is far more epistemically hazardous than evaluating developing world charities. See filter bubbles, tribalism, politically motivated distortions, confirmation bias, etc.) I think the way to do it would be to put an ideologically diverse group of EAs together in the same office reading papers and discussing political issues full-time.

This is similar to a “think tank”, which already exists. I think nonpartisan groups like the RAND Corporation and maybe thoughtful news sources like Vox.com and the Center for Public Integrity are probably the closest thing to EA organizations in the US politics space. They’ve been accumulating prestige & influence for a while. So it’s not clear to me whether the best approach would be to create a rival org or try to influence those orgs to take a more EA approach (through funding them, joining up, etc.) Getting partisan think tanks to work together better (e.g. conduct studies together?) is a more interesting and exotic idea. Howie Lempel might have something to say since he used to work at a think tank.

Maybe this would be worth expanding in to a post on this forum with a more formal & fleshed-out proposal?

Yes, I think these ideas are pretty good - especially working more closely with politically enaged people and deeply engaging people from different ideational political traditions. I'm a little worried that some of these exercises might create the internal controversy and confusion without changing anything? It might be interesting to provide a platform for EAs engaging in political opportunities as they see them as individuals and catch up with the rest of the group about what they're learning? There's a lot you can do before having to transform the nature of political debates...

To push back a bit against the fear of multiple movements, it seems like you could have multiple movements that all overlap with EA. for instance, animal rights and global poverty both overlap with EA, as does immigration work and criminal justice, increasingly. This parallels social justice movements where, say, gay rights and women's rights both overlap with social justice, and I don't see too much harm coming from this (indeed, many social justice movements have tight alliances). Separation might allow specialization, which could easily be net positive.

I've observed the last point happening among many people I know, including myself - people absorb and act on EA values while not necessarily remaining in the community. It seems like creating people who act by EA values is more important than creating an EA community per se. I'm not sure how much community building helps bring this about, though. It would be interesting to see more research and thinking on the connection between community building and generating new EAs (a similar debate is going on among animal activists at the moment).

Where is the Ben post that this is referring to?

Preventative remark: I asked Evan not to link the post in the OP because I was worried it would distract the comments here into debates about the object-level merits of the discussion I was reacting to. I don't think that's a productive direction to go in on an Internet forum (not in general, just unproductive here, specifically because the Internet is a terrible medium for conveying nuance/defusing tension), so I'd suggest that people stay away from that topic here.

(Not that this conversation seemed to be going in that direction; just making this remark out of an abundance of caution.)

Gotcha. If thats the case it'd have relaxed my mind if that point was addressed briefly rather than left to be assumed.