At the most recent Effective Altruism Global in San Francisco, I presented CFAR's Double Crux technique for resolving disagreements.
For the "practice" part of the talk, I handed out a series of prompts on EA topics, to generate disagreements to explore. Several people liked the prompts a lot and asked that I make them generally available.
So here they are! Feel free to use these at EA events, as starting points for discussions or Double Cruxes.
(For Double Crux practice at least, I recommend giving each person time to consider their own answers before talking with their partner. I find that when I don't do this, participants are often influenced by each-other's beliefs and the conversation is less fruitful.)
Prompts:
You have a million dollars to donate to the organization or charity of your choice. To which organization do you give it?
If you could make one change to EA culture on a single dimension (more X, or less Y), what change would you make? (Another framing on this question: what annoys you about EA or EAs?)
Should EA branding ever lie or mislead in some way, in order to get people to donate to effective charities? (For instance: exaggerating the cheapness with which a human life can be saved might get people more excited about donating to the Against Malaria Foundation and similar.)
You’re in charge of outreach for EA. You have to choose one demographic to focus on for introducing EA concepts to, and bringing into the movement. What single demographic do you prioritize?
Are there any causes that should not be included as a part of EA?
Should EA be trying to grow as quickly as possible?
This is the printout that I typically use when teaching Double Crux to EAs, which includes instructions and some additional binary, non-EA questions.
I think the double crux game can be good for dispute resolution. But I think generating disagreement even in a sandbox environment can be counterproductive. It's similar to how having public debates on its face appears seems like it can better resolve a dispute, but if one isn't willing to debate entirely in good faith, they can ruin the debate to the point it shouldn't have happened in the first place. Even if a disagreement isn't socially bad in that it will persist as a conflict after a failed double crux game, it could limit effective altruists to black-and-white thinking after the fact. This lends itself to an absence of the creative problem-solving EA needs.
Perhaps even more than collaborative truth-seeking, the EA community needs individual EAs to learn to think for themselves more to generate possible solutions that the community's core can't solve themselves. There are a lot of EAs who have spare time on their hands that could be better used without something to put it towards. I think starting independent projects an be a valuable use of that time. Here are some of these questions reframed to prompt effective altruists to generate creative solutions.
Imagine you've been given discretion of 10% of the Open Philanthropy Project's annual grantmaking budget. How would you distribute it?
How would solve what you see as the biggest cultural problem in EA?
Under what conditions do you think the EA movement would be justified in deliberately deceiving or misleading the public?
How should EA address our outreach blindspots?
At what rate should EA be growing? How should that be managed?
These questions are reframed to be more challenging. But that's my goal. I think many individual EAs should be challenged to generate less confused models on these topics, and from there between models is when deliberation like double crux should start. Especially if they start from a place of ignorance on current thinking on these issues in EA[1], I don't think in the span of only a couple minutes either side of a double crux game will generate an excellent but controversial hypothesis worth challenging.
The examples in the questions provided are open questions in EA EA organizations don't themselves have good answers to, and I'm sure they'd appreciate additional thinking and support building off their ideas. These aren't binary questions with just one of two possible solutions. I think using EA examples in the double crux game may be a bad idea because it will inadvertently lead EAs to come away with a more simplistic impression of these issues than they should. There is no problem with the double crux game, but maybe EAs should learn it without using EA examples.
[1] This sounds callous, but I think it's a common coordination problem we need to fix. It isn't hard, as it's actually quite easy to miss important theoretical developments that make the rounds among EA orgs but aren't broadcast to the broader movement.
I like these modified questions.
The reason why the original formulations are what they are is to get out of the trap of everyone agreeing that "good things are good", and to draw out specific disagreements.
The intention is that each of these has some sort of crisp "yes or no" or "we should or shouldn't prioritize X". But also the crisp "yes or no" is rooted in a detailed, and potentially original, model.