Hide table of contents

The Centre for Exploratory Altruism Research (CEARCH) is an EA organization working on cause prioritization research as well as grantmaking and donor advisory. This project was commissioned by the leadership of the Meta Charity Funders (MCF) – also known as the Meta Charity Funding Circle (MCFC) – with the objective of identifying what is underfunded vs overfunded in EA meta. The views expressed in this report are CEARCH's and do not necessarily reflect the position of the MCF.

Generally, by meta we refer to projects whose theory of change is indirect, and involving improving the EA movement's ability to do good – for example, via cause/intervention/charity prioritization (i.e. improving our knowledge of what is cost-effective); effective giving (i.e. increasing the pool of money donated in an impact-oriented way); or talent development (i.e. increasing the pool and ability of people willing and able to work in impactful careers).

The full report may be found here (link). Note that the public version of the report is partially redacted, to respect the confidentiality of certain grants, as well as the anonymity of the people whom we interviewed or surveyed.

 

Quantitative Findings

 

Detailed Findings

To access our detailed findings, refer to our spreadsheet (link).

 

Overall Meta Funding

Aggregate EA meta funding saw rapid growth and equally rapid contraction over 2021 to 2023 – growing from 109 million in 2021 to 193 million in 2022, before shrinking back to 117 million in 2023. The analysis excludes FTX, as ongoing clawbacks mean that their funding has functioned less as grants and more as loans.

  • Open Philanthropy is by far the biggest funder in the space, and changes in the meta funding landscape are largely driven by changes in OP's spending. And indeed, OP's global catastrophic risks (GCR) capacity building grants tripled from 2021 to 2022, before falling to twice the 2021 baseline in 2023.
  • This finding is in line with Tyler Maule's previous analysis.

 

Meta Funding by Cause Area

The funding allocation by cause was, in descending order: (most funding) longtermism (i.e. AI, biosecurity, nuclear etc) (274 million) >> global health and development (GHD) (67 million) > cross-cause (53 million) > animal welfare (25 million) (least funding).

 

Meta Funding by Intervention

The funding allocation by intervention was, in descending order: (most funding) other/miscellaneous (e.g. general community building, including by national/local EA organizations; events; community infrastructure; co-working spaces; fellowships for community builders; production of EA-adjacent media content; translation projects; student outreach; and book purchases etc) (193 million) > talent (121 million) > prioritization (92 million) >> effective giving (13 million) (least funding). One note of caution – we believe our results overstate how well funded prioritization is, relative to the other three intervention types.

  • We take into account what grantmakers spend internally on prioritization research, but for lack of time, we do not perform an equivalent analysis for non-grantmakers (i.e. imputing their budget to effective giving, talent, and other/miscellaneous).
  • For simplicity, we classified all grantmakers (except GWWC) as engaging in prioritization, though some grantmakers (e.g. Founders Pledge, Longview) also do effective giving work.

 

Meta Funding by Cause Area & Intervention

The funding allocation by cause/intervention subgroup was as follows:

  • The areas with the most funding were: longtermist other/miscellaneous (153 million) > longtermist talent (106 million) > GHD prioritization (51 million).
  • The areas with the least funding were GHD other/miscellaneous (3,000,000) > animal welfare effective giving (500,000) > longtermist effective giving (100,000).


 

Meta Funding by Organization & Size Ranking

Open Philanthropy is by far the largest funder in the meta space, having given 299 million between 2021-2023; the next two biggest funders are GiveWell (41 million), which also receives OP funding, and EA Funds (which is also funded by OP to a very significant extent).
 

Meta Funding by Cause Area & Organization

  • OP has allocated the lion's share of its meta funding towards longtermism.
  • EA Funds, MCF, GWWC and CE have prioritized cross-cause meta.
  • As expected, GiveWell, ACE and SFF naturally focus on GHD, animal welfare and longtermist meta respectively.
  • FP has supported GHD meta, as has HLI.
  • ACX and Manifund have mainly (or exclusively) given to longtermist meta.

 

How Geography Affects Funding

The survey we conducted finds no strong evidence of funding being geographically biased towards the US/UK relative to the rest of the world; however, there is considerable reason to be sceptical of these findings:

  • The small sample size leads to low power, and a low probability of rejecting the null hypothesis even if there is a real relationship between geography and funding success.
  • Selection bias likely in play. Organizations/individuals who have funding are systematically less likely to take the survey (since they are likely to be busy working on their funded projects – in contrast to organizations/individuals who did not get funding, who might find the survey an outlet for airing their dissatisfaction with funders). Hence, even if it were true that US/UK projects are more likely to be funded, this will not show up in the data, due to funded organizations/individuals self-selecting out of the survey pool.
  • Applicants presumably decide to apply for funding based on their perceived chances of success. If US/UK organizations/individuals genuinely have a better chance of securing funding, and knew this fact, then even lower quality projects would choose to apply, thus lowering the average quality of US/UK applications in the pool. Hence, we end up comparing higher quality projects from the rest of the world to lower quality projects from the US/UK, and equal success in fundraising is not evidence that projects of identical quality from different parts of the world have genuinely equal access to funding. 

Overall, we would not update much away from a prior that US/UK-based projects indeed find it easier to get funding due to proximity to grantmakers and donors in the US and UK. To put things in context, the biggest grantmakers on our list are generally based in SF or London (though this is also a generalization to the extent that the teams are global in nature).

 

Core vs Non-Core Funders

Overall, the EA meta grantmaking landscape is highly concentrated, with core funders (i.e. OP, EA Funds & SFF) making up 82% of total funding.

 

Biggest EA Organizations

The funding landscape is also fairly skewed on the recipients' end, with the top 10 most important EA organizations receiving 31% of total funding.

 

Expenditure Categories

When EA organizations spend, they tend to spend approximately 58% of their budget on salary, 8% on non-salary staff costs, and 35% on all other costs.

 

Qualitative Findings

 

Outline

We interviewed numerous experts, including but not limited to staff employed by (or donors associated with) the following organizations: OP, EA Funds, MCF, GiveWell, ACE, SFF, FP, GWWC, CE, HLI and CEA. We also surveyed the EA community at large. We are grateful to everyone involved for their valuable insights and invaluable help.

 

Prioritizing Direct vs Meta

Crucial considerations include (a) the cost-effectiveness of the direct work you yourself would otherwise fund, relative to the cost-effectiveness of the direct work the average new EA donor funds; and (b) whether opportunities are getting better or worse.


 Prioritizing Between Causes

Whether to prioritize animal welfare relative to GHD depends on welfare ranges and the effectiveness of top animal interventions, particularly advocacy, in light of counterfactual attribution of advocacy success and declining marginal returns. Whether to prioritize longtermism relative to GHD and animal welfare turns on philosophical considerations (e.g. the person-affecting view vs totalism) and epistemic considerations (e.g. theoretical reasoning vs empirical evidence).

 

Prioritizing Between Interventions

  • Effective Giving: Giving What We Can does appear to generate more money than they spend. There are legitimate concerns that GWWC's current estimate of their own cost-effectiveness is overly optimistic, but a previous shallow appraisal by CEARCH identified no significant flaws, and we believe that GWWC's giving multiplier is directionally correct (i.e. positive). 

    • Effective giving and the talent pipeline are not necessarily mutually exclusive – GWWC is an important source of talent for EA, and it is counterproductive to reduce effective giving outreach efforts.
       
    • There is a strong case to be made that EA has a backwards approach where we should be focusing EG outreach efforts on high income individuals in rich countries, particularly the US, and yet it is these same people who end up doing a lot of the research and direct work (e.g. in AI), while the people who do EG in practice have a lot less intrinsic earning potential.
       
    • Similarly, it is arguable that younger EAs should be encouraged to do earning-to-give/normal careers where they donate, since one has fewer useful skills and less experience while young, and money is the most useful thing one possesses then (especially in a context where children or mortgages are not a concern). In contrast, EAs can contribute more in direct work when older, when in possession of greater practical skills and experience. The exception to this rule would be EA charity entrepreneurship, where younger people can afford to bear more risk.
       
  • Talent: The flipside to the above discussion is that our talent outreach might be worth refocusing, both geographically and in terms of age. Geographically, India may be a standout opportunity for getting talent to do research/direct work in a counterfactually cheap way. In terms of age prioritization, it is suboptimal that EA focuses more on outreach to university students or young professionals as opposed to mid-career people with greater expertise and experience.
     
  • Prioritization: GHD cause prioritization may be a good investment if a funder has a higher risk tolerance/a greater willingness to accept a hits-based approach (e.g. health policy advocacy) compared to the average GHD funder.
     
  • Community building: This may be a less valuable intervention unless community building organizations have clearer theories of change and externally validated cost-effectiveness analyses that show how their work produces impact and not just engagement.

 

Open Philanthropy & EA Funds

  • Open Philanthropy: (1) The GHW EA team seems likely to continue prioritizing effective giving; (2) the animal welfare team is currently prioritizing a mix of interventions; and (3) the global catastrophic risks (GCR) capacity building team will likely continue prioritizing AGI risk.
  • There has been increasing criticism of OP, stemming from (a) cause level disagreements (neartermists vs longtermists, animal welfare vs GHD); (b) strategic disagreements (e.g. lack of leadership; and spending on GCR and EA community building doing little, or perhaps even accelerating AI risk); and (c) concerns over the dominance of OP warping the community to reduce accountability and impact.
  • EA Funds: The EA Funds chair has clarified that EAIF would only really coordinate with OP, since they're reliably around; only if the MCF was around for some time, would EA Funds find it worth factoring into their plans.
  • On the issue of EA Funds's performance, we found a convergence in opinion amongst both experts and the EA community at large. It was generally thought that (a) EA Funds leadership has not performed too well relative to the complex needs of the ecosystem; and (b) the EAIF in particular has not been operating smoothly. Our own assessment is that the criticism is not without merit, and overall, we are (a) concerned about EAIF's current ability to fully support an effective EA meta ecosystem; and (b) uncertain as to whether this will improve in the medium term. On the positive side, we believe improvement to be possible, and in some respects fairly straightforward (e.g. hiring additional staff to process EAIF grants).

 

  • Trends: Potential trends of note within the funding space include (a) more money being available on the longtermist side due to massive differences in willingness-to-pay between different parts of OP; (b) an increasing trend towards funding effective giving; and (c) much greater interest in funding field- and movement-building outside of EA (e.g. AI safety field-building) 
  • Uncertainty: The meta funding landscape faces high uncertainty in the short-to-medium term, particularly due to (a) turnover at OP (i.e. James Snowden and Claire Zabel moving out of their roles); (b) uncertainty at EA Funds, both in terms of what it wants to do and what funding it has; and (c) the short duration of time since FTX collapsed, with the consequences still unfolding – notably, clawbacks are still occuring.

76

3
9

Reactions

3
9

More posts like this

Comments39
Sorted by Click to highlight new comments since: Today at 10:45 AM
Linch
6d88
12
2
3

(I work for EA Funds, including EAIF, helping out with public communications among other work. I'm not a grantmaker on EAIF and I'm not responsible for any decision on any specific EAIF grant). 

Hi. Thanks for writing this. I appreciate you putting the work in this, even though I strongly disagree with the framing of most of the doc that I feel informed enough to opine on, as well as most of the object-level. 

Ultimately, I think the parts of your report about EA Funds are mostly incorrect or substantively misleading, given the best information I have available. But I think it’s possible I’m misunderstanding your position or I don’t have enough context. So please read the following as my own best understanding of the situation, which can definitely be wrong. But first, onto the positives:

  • I appreciate that the critical points in the doc are made as technical critiques, rather than paradigmatic ones. Technical critiques are ones that people are actually compelled to respond to, and can actually compel action (rather than just make people feel vaguely bad/smug and don’t compel any change).
  • The report has many numerical/quantitative details. In theory, those are easier to falsify.
  • The report appears extensive and must have taken a long time to write.

There are also some things the report mentioned that we have also been tracking, and I believe we have substantial room for improvement:

  • Our grant evaluation process is still slower than we would like.
    • While broadly faster than other comparable funds I’m aware of (see this comment), I still think we have substantial room for improvement.
  • Our various subfunds, particularly EAIF, have at points been understaffed and under-capacity. 
    • While I strongly disagree with the assessment that hiring more grantmakers is “fairly straightforward” (bad calls with grant evaluations are very costly, for reasons including but not limited to insufficient attention to adversarial selection; empirically most EA grantmaking organizations have found it very difficult to hire), I do think on the margin we can do significantly more on hiring.
  • Our limited capacity has made it difficult for us to communicate with and/or coordinate with all the other stakeholders in the system, so we’re probably missing out on key high-EV opportunities (eg several of our existing collaborations in AI safety have started later than they counterfactually could have, and we haven’t been able to schedule time to fly to coordinate with folks in London/DC/Netherlands/Sweden.
    • One of the reasons I came on to EA Funds full-time is to help communicate with various groups. 

Now, onto the disagreements:

Procedurally:

  • I was surprised that so much of the views that were ascribed to “EA Funds’ leadership” were from notes taken from a single informal call with the EA Funds project lead, that you did not confirm was okay to be shared publicly. 
    • They said that they usually explicitly request privacy before quoting them publicly, but are not sure if they did so in this instance.
    • They were also surprised that there was a public report out at all.
      • My best guess is that there was a misunderstanding that arose from a norm difference, where you come from the expectation that professional meetings are expected to be public unless explicitly stated otherwise, whereas the norm that I (and I think most EAs?) are more used to is that 1-1 meetings are private unless explicitly stated otherwise. 
    • They also disagree with the characterization of almost all of their comments (will say more below), which I think speaks to the epistemic advantages of confirming before publicly attributing comments made by someone else.
  • I’d have found it helpful if you shared a copy of the post before making it public.
    • We could’ve corrected most misunderstandings
    • If you were too busy for private corrections, I could’ve at least written this response earlier.
  • Many things in general were false (more details below). Having a fact-checking process might be useful going forwards.

 

Substantively:

  • When the report quoted “CEA has had to step in and provide support in evaluating EAIF grants for them” I believe this is false or at least substantively misleading.
    • The closest thing I can think about is that we ask CEA Comm Health for help in reviewing comm health issues with our grants (which as I understand is part of their explicit job duties and both sides are happy with)”
      • It’s possible your source misunderstood the relationship and thought the Comm Health work was supererogatory or accidental? 
    • We frequently ask technical advisors for advice on project direction in a less institutionalized capacity as well[1]. I think “step in” conveys the wrong understanding, as most of the grant evaluation is still done by the various funds.
    • (To confirm this impression, we checked with multiple senior people involved with CEA’s work; however, this was not an exhaustive sweep and it’s definitely possible that my impression is incorrect). 
  • While it is true that we’re much slower than we would like, it seems very unreasonable to single out EA Funds grantmaking as "unreasonably long" when other grantmakers are as slow or slower. 
    • See e.g. Abraham Rowe’s comment here
      • “"Not weighing in on LTFF specifically, but from having done a lot of traditional nonprofit fundraising, I'd guess two months is a faster response time than 80% of foundations/institutional funders, and one month is probably faster than like 95%+. My best guess at the average for traditional nonprofit funders is more like 3-6 months. I guess my impression is that even in the worst cases, EA Funds has been operating pretty well above average compared to the traditional nonprofit funding world (though perhaps that isn't the right comparison). Given that LTFF is funding a lot of research, 2 months is almost certainly better than most academic grants.
      • My impression from what I think is a pretty large sample of EA funders and grants is also that EA Funds is the fastest turnaround time on average compared to the list you mention [Editor's note: "Open Phil, SFF, Founders Pledge, and Longview"] ([with] exceptions in some cases in both directions for EA Funds and other funders)”
      • I also sanity-checked with both Google Search and GPT-4[2]
    • Broadly I’m aware that other people on the forum also believes that we’re slow, but I think most people who believe this believe so because:
      • We have our own aspirations to be faster, and we try to do so.
      • They think from a first-principles perspective that grantmakers “can” be faster.
      • We talk about our decisions and process very publicly, and so become more of an easy target for applicants’ grievances.
    • But while I understand and sympathize with other people’s frustrations[3], it is probably not factually true that we’re slower in relative terms than other organizations, and it’s odd to single out EA Funds here.
  • When your report said, “Critically, the expert reports that another major meta donor found EA Funds leadership frustrating to work with, and so ended up disengaging from further meta grantmaking coordination” my guess is that the quoted position is not true.
    • I feel like we’re keeping tabs on all the major donors (OP, SFF, Longview, etc).So I’m not sure who they could possibly be referring to.
      • Though I guess it’s possible that there is a major donor that’s so annoyed with us that they made efforts to hide themselves from us so we haven’t even heard about them.
      • But I think it’s more likely that the person in question isn’t a major donor.
    • To be clear, any donor feeling frustrated with us is regrettable. While it is true that not all donors can or should want to work with us (e.g. due to sufficiently differing cause prioritization or empirical worldviews), it is still regrettable that people have an emotionally frustrating experience. 
  • Your report says that EA Funds leadership was “strongly dismissing the value of prioritization research, where other grantmakers generally expressed higher uncertainty”, but this is false.
    • I want to be clear that EA Funds has historically been, and currently is, quite positive on cause prioritization in general (though of course specific work may be lower quality, or good work that’s not cause prioritization may be falsely labeled as cause prioritization).
    • By revealed preferences, EAIF has given very large grants to worldview investigations and moral weights work at Rethink Priorities
    • By stated preferences, “research that aids prioritization across different cause areas” was listed as one of the central examples of things that EAIF would be excited to fund.
    • My understanding is that the best evidence you have for this view is that EA Funds leadership “would consider less than half of what [Rethink Priorities] does cause prioritization.” 
    • I’m confused why you think the quoted statement is surprising or good evidence, given that the stated claim is just obviously true. Eg, eg The Rodenticide Reduction Sequence or Cultured meat: A comparison of techno-economic analyses or Exposure to Lead Paint in Low- and Middle-Income Countries (To give three examples of work that I have more than a passing familiarity with) are much more about intervention prioritization than intercause prioritization. An example of the latter is the moral weights work at Rethink Priorities Worldview Investigations.
    • My best guess is that this is just a semantics misunderstanding, where EA Funds’ project lead was trying to convey a technical point about the difference between intercause cause prioritization vs intervention prioritization, whereas you understood his claim as a emotive position of “boo cause prioritization”
  • Your report states that “EA Funds leadership doesn't believe that there is more uncertainty now with EA Fund's funding compared to other points in time” This is clearly false. 
    • I knew coming on to EA Funds that the job will have greater job uncertainty than other jobs I’ve had in the past, and I believe this was adequately communicated to me. 
    • We think about funding uncertainty a lot. EA Funds’ funding has always been uncertain, and things have gotten worse since Nov 2022.
    • Nor would it be consistent with either our stated or revealed preferences. 
      • Our revealed preference is that we spend substantially more staff time on fundraising than we have in the past.
      • Our stated preferences include “I generally expect our funding bar to vary more over time and to depend more on individual donations than it has historically.” and “LTFF and EAIF are unusually funding-constrained right now” 
        • The last one was even the title of a post with 4000+ views!
        • I don’t have a good idea for how much more unambiguous we could be.

 

Semantically:

I originally want to correct misunderstandings and misrepresentations of EA Funds’ positions more broadly in the report. However I think there were just a lot of misunderstandings overall, so I think it's simpler for people to just assume I contest almost every categorization of the form “EA funds believes X”. A few select examples:

  • When your report claimed “​​leadership is of the view that the current funding landscape isn't more difficult for community builders” I a) don’t think we’ve said that, and b) to the extent we believe it, it’s relative to eg 2019; it’d be false compared to the 2022 era of unjustly excessive spending.
  • “The EA Funds chair has clarified that EAIF would only really coordinate with OP, since they're reliably around; only if the [Meta-Charity Funders] was around for some time, would EA Funds find it worth factoring into their plans. ” 
    • To clarify, at the time the fund chair was unsure if MCF was only going to have one round. If they only have one round, it wouldn't make sense to change EA Funds’ strategy based on that. If they have multiple rounds (eg, more than 2, it could be worth factoring in). The costs of coordination are nontrivially significant.
    • It’s also worth noting that the fund chair had 2 calls with MCF and passed on various grants that they thought MCF might be interested in evaluating, which some people may consider coordination.
    • We’ve also coordinated more extensively with other non-OP funders, and have plans in the works for other collaborations with large funders.
  • “In general, they don't think that other funders outside of OP need to do work on prioritization, and are in general sceptical of such work. ”
    • AFAIK nobody at EA Funds believes this.
  • [EA funds believes] “so if EA groups struggle to raise money, it's simply because there are more compelling opportunities available instead.”
    • The statement seems kind of conceptually confused? Funders should always be trying to give to the most cost-effective projects on the margin.
    • The most charitable position that’s similar to the above I could think of is that some people might believe that 
      • “most community-building projects that are not funded now aren’t funded because of constraints on grantmaker capacity,” so grantmakers make poor decisions
        • Note that the corollary to the above is that many of the community building projects that are funded should not have been funded. 
      • I can’t speak to other people at EA Funds, but my own best guess is that this is not true (for projects that people online are likely to have heard of).
        • I’m more optimistic about boutique funding arrangements for projects within people’s networks that are unlikely to have applied to big funders, or people inspiring those around them to create new projects.
    • If projects aren’t funded, the biggest high-level reason is that there are limited resources in the world in general, and in EA specifically. You might additionally also believe that meta-EA in general is underfunded relative to object-level programs. 
  • (Minor): To clarify “since OP's GHW EA team is focusing on effective giving, EAIF will consider this less neglected” we should caveat this by saying that less neglected doesn’t necessarily mean less cost-effective than other plausible things for EAIF to fund.
  • Re: “EA Funds not posting reports or having public metrics of success”
    • We do post public payout reports.
    • My understanding is that you are (understandably) upset that we don’t have clear metrics and cost-effectiveness analyses written up. 
      • I think this is a reasonable and understandable complaint, and we have indeed gotten this feedback before from others and have substantial room to improve here.
    • However, I think many readers might interpret the statement as something stronger, e.g. interpreting it as us not posting reports or writing publicly much at all
      • As a matter of practice, we write a lot more about what we fund and our decision process than any other EA funder I’m aware of (and likely more than most other non-EA funders). I think many readers may get the wrong impression of our level of transparency from that comment.

Note to readers: I reached out to Joel to clarify some of these points before posting. I really appreciate his prompt responses! Due to time constraints, I decided to not send him a copy of this exact comment before posting publicly.

  1. ^

    I personally have benefited greatly from talking to specialist advisors in biosecurity.

  2. ^

    From GPT4

    “The median time to receive a response for an academic grant can vary significantly depending on the funding organization, the field of study, and the specific grant program. Generally, the process can take anywhere from a few months to over a year. ”
    “The timeline for receiving a response on grant applications can vary across different fields and types of grants, but generally, the processes are similar in length to those in the academic and scientific research sectors.”
    “Smaller grants in this field might be decided upon quicker, potentially within 3 to 6 months [emphasis mine], especially if they require less funding or involve fewer regulatory hurdles.”

  3. ^

    Being funded by grants kind of sucks as an experience compared to e.g. employment; I dislike adding to such frustrations. There are also several cases I’m aware of where counterfactually impactful projects were not taken due to funders being insufficiently able to fund things in time, in some of those incidences I'm more responsible than anybody else.

Hi Linch,

Thanks for engaging. I appreciate that we can have a fairly object-level disagreement over this issue; it's not personal, one way or another.

Meta point to start: We do not make any of these criticisms of EA Funds lightly, and when we do, it's against our own interests, because we ourselves are potentially dependent on EAIF for future funding.

To address the points brought up, generally in the order that you raised them:

(1) On the fundamental matter of publication. I would like to flag out that, from checking the email chain plus our own conversation notes (both verbatim and cleaned-up), there was no request that this not be publicized.

For all our interviews, whenever someone flagged out that X data or Y document or indeed the conversation in general shouldn't be publicized, we respected this and did not do so. In the public version of the report, this is most evident in our spreadsheet where a whole bunch of grant details have been redacted; but more generally, anyone with the "true" version of the report shared with the MCF leadership will also be able to spot differences. We also redacted all qualitative feedback from the community survey, and by default anonymized all expert interviewees who gave criticisms of large grantmakers, to protect them from backlash. 

I would also note that we generally attributed views to, and discussed, "EA Leadership" in the abstract, both because we didn't want to make this a personal criticism, and also because it afforded a degree of anonymity.

At the end of the day, I apologize if the publication was not in line with what EA Funds would have wanted - I agree it's probably a difference in norms. In a professional context, I'm generally comfortable with people relaying that I said X in private, unless there was an explicit request not to share (e.g. I was talking to a UK-based donor yesterday, and I shared a bunch of my grantmaking views. If he wrote a post on the forum summarizing the conversations he had with a bunch of research organizations and donor advisory orgs, including our own, I wouldn't object). More generally, I think if we have some degree of public influence (including by the money we control) it would be difficult from the perspective of public accountability if "insiders" such as ourselves were unwilling to share with the public what we think or know.

(2) For the issue of CEA stepping in: In our previous conversation, you relayed that you asked a senior person at CEA and they in turn said that "they’re aware of some things that might make the statement technically true but misleading, and they are not aware of anything that would make the statement non-misleading, although this isn’t authoritative since many thing happened at CEA". For the record, I'm happy to remove this since the help/assistance, if any, doesn't seem too material one way or another.

(3) For whether it's fair to characterize EAIF's grant timelines as unreasonably long. As previously discussed, I think the relevant metric is EAIF's own declared timetable ("The Animal Welfare Fund, Long-Term Future Fund and EA Infrastructure Fund aim to respond to all applications in 2 months and most applications in 3 weeks."). This is because organizations and individuals make plans based on when they expect to get an answer - when to begin applying; whether to start or stop projects; whether to go find another job; whether to hire or fire; whether to reach out to another grantmaker who isn't going to support you until and unless you have already exhausted the primary avenues of potential funding.

(4) The issue of the major donor we relayed was frustrated/turned off. You flag out that you're keeping tabs on all the major donors, and so don't think the person in question is major. While I agree that it's somewhat subjective - it's also true that this is a HNWI who, beyond their own giving, is also sitting on the legal or advisory boards many other significant grantmakers and philanthropic outfits. Also, knowledgeable EAs in the space have generally characterized this person as an important meta funder to me (in the context of my own organization then thinking of fundraising, and being advised as to whom to approach). So even if they aren't major in the sense that OP (or EA Funds are), they could reasonably be considered fairly significant. In any case, the discussion is backwards, I think - I agree that they don't play as significant a role in the community right now (and so you assessment of them as non-major is reasonable), but that would be because of the frustration they have had with EA Funds (and, to be fair, the EA community in general, I understand). So perhaps it's best to understand this as potentially vs currently major.

(5) On whether it's fair to characterize EA Funds leadership as being strongly dismissive of cause prioritization. We agree that grants have been made to RP; so the question is cause prioritization outside OP and OP-funded RP. Our assessment of EA Fund's general scepticism of prioritization was based, among other things, on what we reported in the previous section "They believe cause prioritization is an area that is talent constrained, and there aren't a lot of people they feel great giving to, and it's not clear what their natural pay would be. They do not think of RP as doing cause prioritization, and though in their view RP could absorb more people/money in a moderately cost-effective way, they would consider less than half of what they do cause prioritization. In general, they don't think that other funders outside of OP need to do work on prioritization, and are in general sceptical of such work." In your comment, you dispute that the bolded part in particular is true, saying "AFAIK nobody at EA Funds believes this."

We have both verbatim and cleaned up/organized notes on this (n.b. we shared both with you privately). So it appears we have a fundamental disagreement here (and also elsewhere) as to whether what we noted down/transcribed is an accurate record of what was actually said.

TLDR: Fundamentally, I stand by the accuracy of our conversation notes.

(a) Epistemically, it's more likely that one doesn't remember what one said previously vs the interviewer (if in good faith) catastrophically misunderstanding and recording down something that wholesale wasn't said at all (as opposed to a more minor error - we agree that that can totally happen; see below)

(b) From my own personal perspective - I used to work in government and in consulting (for governments). It was standard practice to have notes of meeting, as made by junior staffers and then submitted to more senior staff for edits and approval. Nothing resembling this happened to either me or anyone else (i.e. just total misunderstanding tantamount to fabrication, in saying that that XYZ was said when nothing of the sort took place).

(c) My word does not need to be taken for this. We interviewed other people, and I'm beginning to reach out to them again to check that our notes matched what they said. One has already responded (the person we labelled Expert 5 on Page 34 of the report); they said "This is all broadly correct" but requested we made some minor edits to the following paragraphs (changes indicated by bold and strikethrough)

  • Expert 5: Reports both substantive and communications-related concerns about EA Funds leadership.

    For the latter, the expert reports both himself and others finding communications with EA Funds leadership difficult and the conversations confusing.

    For the substantive concerns – beyond the long wait times EAIF imposes on grantees, the expert was primarily worried that EA Funds leadership has been unreceptive to new ideas and that they are unjustifiably confident that EA Funds is fundamentally correct in its grantmaking decisions. In particular, it appears to the expert that EA Funds leadership does not believe that additional sources of meta funding would be useful for non-EAIF grants [phrase added] – they believe that projects unfunded by EAIF do not deserve funding at all (rather than some projects perhaps not being the right fit for the EAIF, but potentially worth funding by other funders with different ethical worldviews, risk aversion or epistemics). Critically, the expert reports that another major meta donor found EA Funds leadership frustrating to work with, and so ended up disengaging from further meta grantmaking coordination and this likely is one reason they ended up disengagement from further meta grantmaking coordination [replaced].

My even handed interpretation of this overall situation (trying to be generous to everyone) is that what was reported here ("In general, they don't think that other funders outside of OP need to do work on prioritization") was something the EA Funds interviewee said relatively casually (not necessarily a deep and abiding view, and so not something worth remembering) - perhaps indicative of scepticism of a lot of cause prioritization work but not literally thinking nothing outside OP/RP is worth funding. (We actually do agree with this scepticism, up to an extent).

(6) On whether our statement that “EA Funds leadership doesn't believe that there is more uncertainty now with EA Fund's funding compared to other points in time” is accurate. You say that this is clearly false. Again, I stand by the accuracy of our conversation notes. And in fact, I actually do personally and distinctively remember this particular exchange, because it stood out, as did the exchange that immediately followed, on whether OP's use of the fund-matching mechanism creates more uncertainty.

My generous interpretation of this situation is, again, some things may be said relatively casually, but may not be indicative of deep, abiding views.

(8) For the various semantic disagreements. Some of it we discussed above (e.g. the OP cause prioritization stuff); for the rest -

On whether this part is accurate: “​​Leadership is of the view that the current funding landscape isn't more difficult for community builders”. Again, we do hold that this was said, based on the transcripts. And again, to be even handed, I think your interpretation (b) is right - probably your team is thinking of the baseline as 2019, while we were thinking mainly of 2021-now.

On whether this part is accurate: “The EA Funds chair has clarified that EAIF would only really coordinate with OP, since they're reliably around; only if the [Meta-Charity Funders] was around for some time, would EA Funds find it worth factoring into their plans. ” I don't think we disagree too much, if we agree that EA Fund's position is that coordination is only worthwhile if the counterpart is around for a bit. Otherwise, it's just some subjective disagreement on what what coordination is or what significant degrees of it amount to.

On this statement: "[EA funds believes] “so if EA groups struggle to raise money, it's simply because there are more compelling opportunities available instead.

In our discussion, I asked about the community building funding landscape being worse; the interviewee disagreed with this characterization, and started discussing how it's more that standards have risen (which we agree is a factor). The issue is that the other factor of objectively less funding being available was not brought up, even though it is, in our view, the dominant factor (and if you asked community builders this will be all they talk about). I think our disagreement here is partly subjective - over what a bad funding landscape is, and also the right degree of emphasis to put on rising standards vs less funding.

(9) EA Funds not posting reports or having public metrics of successes. Per our internal back-and-forth, we've clarified that we mean reports of success or having public metrics of success. We didn't view reports on payouts to be evidence of success, since payouts are a cost, and not the desired end goal in itself. This contrasts with reports on output (e.g. a community building grant actually leading to increased engagement on XYZ engagement metrics) or much more preferably, report on impact (e.g. and those XYZ engagement metrics leading to actual money donated to GiveWell, from which we can infer that X lives were saved). Like, speaking for my own organization, I don't think the people funding our regranting budgets would be happy if I reported the mere spending as evidence of success.

(OVERALL) For what it's worth, I'm happy to agree to disagree, and call it a day. Both your team and mine are busy with our actual work of research/grantmaking/etc, and I'm not sure if further back and forth will be particularly productive, or a good use of my time or yours.

Lizka
5d33
6
1
1

I'm going to butt in with some quick comments, mostly because:

  • I think it's pretty important to make sure the report isn't causing serious misunderstandings 
  • and because I think it can be quite stressful for people to respond to (potentially incorrect) criticisms of their projects — or to content that seem to misrepresent their project(s) — and I think it can help if someone else helps disentangle/clarify things a bit. (To be clear, I haven't run this past Linch and don't know if he's actually finding this stressful or the like. And I don't want to discourage critical content or suggest that it's inherently harmful; I just think external people can help in this kind of discussion.)

I'm sharing comments and suggestions below, using your (Joel's) numbering. (In general, I'm not sharing my overall views on EA Funds or the report. I'm just trying to clarify some confusions that seem resolvable, based on the above discussion, and suggest changes that I hope would make the report more useful.)

  • (2) Given that apparently the claim that "CEA has had to step in and provide support" EA Funds is likely "technically misleading", it seems good to in fact remove it from the report (or keep it in but immediately and explicitly flag that this seems likely misleading and link Linch's comment) — you said you're happy to do this, and I'd be glad to see it actually removed. 
  • (3) The report currently concludes that would-be grantees "wait an unreasonable amount of time before knowing their grant application results." Linch points out that other grantmakers tend to have similar or longer timelines, and you don't seem to disagree (but argue that it's important to compare the timelines to what EA Funds sets as the expectation for applicants, instead of comparing them to other grantmakers' timelines). 
    • Given that, I'd suggest replacing "unreasonably long" (which implies a criticism of the length itself) with something like "longer than what the website/communications with applicants suggest" (which seems like what you actually believe) everywhere in the report. 
  • (9) The report currently states (or suggests) that EA Funds doesn't post reports publicly. Linch points out that they "do post public payout reports." It seems like you're mostly disagreeing about the kind of reports that should be shared.[3] 
    • Given that this is the case, I think you should clarify this in the report (which currently seems to mislead readers into believing that EA Funds doesn't actually post any public reports), e.g. by replacing "EA Funds [doesn't post] reports or [have] public metrics of success" with "EA Funds posts public payout reports like this, but doesn't have public reports about successes achieved by their grantees." 
  • (5), (6), (8) (and (1)) There are a bunch of disagreements about whether what's described as views of "EA Funds leadership" in the report is an accurate representation of the views.
    • (1) In general, Linch — who has first-hand knowledge — points out that these positions are from "notes taken from a single informal call with the EA Funds project lead" and that the person in question disagrees with "the characterization of almost all of their comments." (Apparently the phrase "EA Funds leadership" was used to avoid criticizing someone personally and to preserve anonymity.)
      • You refer to the notes a lot, explaining that the views in the report are backed by the notes from the call and arguing that one should generally trust notes like this more than someone's recollection of a conversation.[1] Whether or not the notes are more accurate than the project lead's recollection of the call, it seems pretty odd to view the notes as a stronger authority on the views of EA Funds than what someone from EA Funds is explicitly saying now, personally and explicitly. (I.e. what matters is whether a statement is true, not whether it was said in a call.) 
        • You might think that (A) Linch is mistaken about what the project lead thinks (in which case I think the project lead will probably clarify), or (B) that (some?) people at EA Funds have views that they disclosed in the call (maybe because the call was informal and they were more open with their views) but are trying to hide or cover up now — or that what was said in the call is indirect evidence for the views (that are now being disavowed). If (B) is what you believe, I think you should be explicit about that. If not, I think you should basically defer to Linch here. 
      • As a general rule, I suggest at least replacing any instance of "EA Funds leadership [believes]" with something like "our notes from a call with someone involved in running EA Funds imply that they think..." and linking Linch's comment for a counterpoint. 
    • Specific examples: 
      • (5) Seems like Linch explicitly disagrees with the idea that EA Funds dismisses the value of prioritization research, and points out that EAIF has given large grants to relevant work from Rethink Priorities. 
        • Given this, I think you should rewrite statements in the report that are misleading. I also think you should probably clarify that EA Funds has given funding to Rethink Priorities.[2]
        • Also, I'm not as confident here, but it might be good to flag the potential for ~unconscious bias in the discussions of the value of cause prio research (due to the fact that CEARCH is working on cause prioritization research). 
      • (6) Whatever was said in the conversation notes, it seems that EA Funds [leadership] does in fact believe that "there is more uncertainty now with [their] funding compared to other points in time." Seems like this should be corrected in the report.
      • (8) Again, what matters isn't what was said, but what is true (and whether the report is misleading about the truth). Linch seems to think that e.g. the statement about coordination is misleading.

I also want to say that I appreciate the work that has gone into the report and got value from e.g. the breakdown of quantitative data about funding — thanks for putting that together. 

And I want to note potential COIs: I'm at CEA (although to be clear I don't know if people at CEA agree with my comment here), briefly helped evaluate LTFF grants in early 2022, and Linch was my manager when I was a fellow at Rethink Priorities in 2021. 

  1. ^

    E.g. 

    We have both verbatim and cleaned up/organized notes on this (n.b. we shared both with you privately). So it appears we have a fundamental disagreement here (and also elsewhere) as to whether what we noted down/transcribed is an accurate record of what was actually said.

    TLDR: Fundamentally, I stand by the accuracy of our conversation notes.

    Epistemically, it's more likely that one doesn't remember what one said previously vs the interviewer (if in good faith) catastrophically misunderstanding and recording down something that wholesale wasn't said at all (as opposed to a more minor error - we agree that that can totally happen; see below) ...

  2. ^

    In relation to this claim: "They do not think of RP as doing cause prioritization, and though in their view RP could absorb more people/money in a moderately cost-effective way, they would consider less than half of what they do cause prioritization."

  3. ^

    "...we mean reports of success or having public metrics of success. We didn't view reports on payouts to be evidence of success, since payouts are a cost, and not the desired end goal in itself. This contrasts with reports on output (e.g. a community building grant actually leading to increased engagement on XYZ engagement metrics) or much more preferably, report on impact (e.g. and those XYZ engagement metrics leading to actual money donated to GiveWell, from which we can infer that X lives were saved)."

Thanks for the clarifications, Joel.

At the end of the day, I apologize if the publication was not in line with what EA Funds would have wanted - I agree it's probably a difference in norms. In a professional context, I'm generally comfortable with people relaying that I said X in private, unless there was an explicit request not to share

It is clear from Linch's comment that he would have liked to see a draft of the report before it was published. Did you underestimate the interest of EA Funds in reviewing the report before its publication, or did you think their interest in reviewing the report was not too relevant? I hope the former.

AGB
9d96
15
7
6
2

Meta note: I wouldn’t normally write a comment like this. I don’t seriously consider 99.99% of charities when making my donations; why single out one? I’m writing anyway because comments so far are not engaging with my perspective, and I hope more detail can help 80,000 hours themselves and others engage better if they wish to do so. As I note at the end, they may quite reasonably not wish to do so.

For background, I was one of the people interviewed for this report, and in 2014-2018 my wife and I were one of 80,000 hours’ largest donors. In recent years it has not made my shortlist of donation options. The report’s characterisation of them - spending a huge amount while not clearly being >0 on the margin - is fairly close to my own view, though clearly I was not the only person to express it. All views expressed below are my own.

I think it is very clear that 80,000 hours have had a tremendous influence on the EA community. I cannot recall anyone stating otherwise, so references to things like the EA survey are not very relevant. But influence is not impact. I commonly hear two views for why this influence may not translate into positive impact:

-80,000 hours prioritises AI well above other cause areas. As a result they commonly push people off paths which are high-impact per other worldviews. So if you disagree with them about AI, you’re going to read things like their case studies and be pretty nonplussed. You’re also likely to have friends who have left very promising career paths because they were told they would do even more good in AI safety. This is my own position.

-80,000 hours is likely more responsible than any other single org for the many EA-influenced people working on AI capabilities. Many of the people who consider AI top priority are negative on this and thus on the org as a whole. This is not my own position, but I mention it because I think it helps explain why (some) people who are very pro-AI may decline to fund.

I suspect this unusual convergence may be why they got singled out; pretty much every meta org has funders skeptical of them for cause prioritisation reasons, but here there are many skeptics in the crowd broadly aligned on prioritisation.

Looping back to my own position, I would offer two ‘fake’ illustrative anecdotes:

Alice read Doing Good Better and was convinced of the merits of donating a moderate fraction of her income to effective charities. Later, she came across 80,000 hours and was convinced by their argument that her career was far more important. However, she found herself unable to take any of the recommended positions. As a result she neither donates nor works in what they would consider a high-impact role; it’s as if neither interaction had ever occurred, except perhaps she feels a bit down about her apparent uselessness.

Bob was having impact in a cause many EAs consider a top priority. But he is epistemically modest, and inclined to defer to the apparent EA consensus- communicated via 80,000 hours - that AI was more important. He switched careers and did find a role with solid - but worse - personal fit. The role is well-paid and engaging day-to-day; Bob sees little reason to reconsider the trade-off, especially since ChatGPT seems to have vindicated 80,000 hours’ prior belief that AI was going to be a big deal. But if pressed he would readily acknowledge that it’s not clear how his work actually improves things. In line with his broad policy on epistemics, he points out the EA leadership is very positive on his approach; who is he to disagree?

Alice and Bob have always been possible problems from my perspective. But in recent years I’ve met far more of them than I did when I was funding 80,000 hours. My circles could certainly be skewed here, but when there’s a lack of good data my approach to such situations is to base my own decisions on my own observations. If my circles are skewed, other people who are seeing very little of Alice and Bob can always choose to fund.

On that last note, I want to reiterate that I cannot think of a single org, meta or otherwise, that does not have its detractors. I suspect there may be some latent belief that an org as central as 80,000 hours has solid support across most EA funders. To the best of my knowledge this is not and has never been the case, for them or for anyone else. I do not think they should aim for that outcome, and I would encourage readers to update ~0 on learning such.

Just want to say here (since I work at 80k & commented abt our impact metrics & other concerns below) that I think it's totally reasonable to:

  1. Disagree with 80,000 Hours's views on AI safety being so high priority, in which case you'll disagree with a big chunk of the organisation's strategy.
  2. Disagree with 80k's views on working in AI companies (which, tl;dr, is that it's complicated and depends on the role and your own situation but is sometimes a good idea). I personally worry about this one a lot and think it really is possible we could be wrong here. It's not obvious what the best thing to do here is, and we discuss this a bunch internally. But we think there's risk in any approach to issue, and are going with our best guess based on talking to people in the field. (We reported on some of their views, some of which were basically 'no don't do it!', here.)
  3. Think that people should prioritise personal fit more than 80k causes them to. To be clear, we think (& 80k's content emphasises) that personal fit matters a lot. But it's possible we don't push this hard enough. Also, because we think it's not the only thing that matters for impact (& so also talk a lot about cause and intervention choice), we tend to present this as a set of considerations to navigate that involves some trade-offs. So It's reasonable to think that 80k encourages too much trading off of personal fit, at least for some people. 
AGB
9d50
7
4
1
1

Thanks Arden. I suspect you don't disagree with the people interviewed for this report all that much then, though ultimately I can only speak for myself. 

One possible disagreement that you and other commenters brought up that which I meant to respond to in my first comment, but forgot: I would not describe 80,000 hours as cause-neutral, as you try to do here and here. This seems to be an empirical disagreement, quoting from second link:

We are cause neutral[1] – we prioritise x-risk reduction because we think it's most pressing, but it’s possible we could learn more that would make us change our priorities.

I don't think that's how it would go. If an individual 80,000 hours member learned things that cause them to downshift their x-risk or AI safety priority, I expect them to leave the org, not for the org to change. Similar observations on hiring. So while all the individuals involved may be cause neutral and open to change in the sense you describe, 80,000 hours itself is not, practically speaking. It's very common for orgs to be more 'sticky' than their constituent employees in this way. 

I appreciate it's a weekend, and you should feel free to take your time to respond to this if indeed you respond at all. Sorry for missing it in the first round. 

Speaking in a personal capacity here --

We do try to be open to changing our minds so that we can be cause neutral in the relevant sense, and we do change our cause rankings periodically and spend time and resources thinking about them (in fact we’re in the middle of thinking through some changes now). But how well set up are we, institutionally, to be able to in practice make changes as big as deprioritising risks from AI if we get good reasons to? I think this is a good question, and want to think about it more. So thanks!

I think it is very clear that 80,000 hours have had a tremendous influence on the EA community... so references to things like the EA survey are not very relevant. But influence is not impact... 80,000 hours prioritises AI well above other cause areas. As a result they commonly push people off paths which are high-impact per other worldviews.

 

Many of the things the EA Survey shows 80,000 Hours doing (e.g. introducing people to EA in the first place, helping people get more involved with EA, making people more likely to remain engaged with EA, introducing people to ideas and contacts that they think are important for their impact, helping people (by their own lights) have more impact), are things which supporters of a wide variety of worldviews and cause areas could view as valuable. Our data suggests that it is not only people who prioritise longtermist causes who are report these benefits from 80,000 Hours. 

Hi David,

There is a significant overlap between EA and AI safety, and it is often unclear whether people supposedly working on AI safety are increasing/decreasing AI risk. So I think it would be helpful if you could point to some (recent) data on how many people are being introduced to global health and development, and animal welfare via 80,000 Hours.

Thanks Vasco.

We actually have a post on this forthcoming, but I can give you the figures for 80,000 Hours specifically now. 

  • Global Poverty: 
    • 80,000 Hours: 
      • 37.3% rated Global Poverty 4/5
      • 28.6% rated Global Poverty 5/5 
    • All respondents:
      • 39.3% rated Global Poverty 4/5
      • 28.7% rated Global Poverty 5/5
    • So the difference is minimal, but this also neglects the fact that the scale of 80K's recruitment swamps any differences in % supporting different causes. Since 80K recruits so many, it is still the second highest source of people who rate Global Poverty most highly (12.5% of such people) after only personal contacts.
  • Animal Welfare:
    • 80,000 Hours:
      • 26.0% rated Animal Welfare 4/5
      • 9.6% rated Animal Welfare 5/5
    • All respondents:
      • 30.1% rated Animal Welfare 4/5
      • 11.8% rated Animal Welfare 5/5
    • Here the difference is slightly bigger, though 80,000 Hours remains among the top sources of recruits rating Animal Welfare highest (specifically, after personal contact, the top sources are 'Other' (11.2%), 'Book, article or blog' (10.7%), 80,000 Hours (9%).

Thanks, David! Strongly upvoted.

To clarify, are those numbers relative to the people who got to know about EA in 2023 (via 80,000 Hours or any source)?

Thanks Vasco!

These are numbers from the most recent full EA Survey (end of 2022), but they're not limited only to people who joined EA in the most recent year. Slicing it by individual cohorts would reduce the sample size a lot. 

My guess is that it would also increase the support for neartermist causes among all recruits (respondents tend to start out neartermist and become more longtermist over time).

That said, if we do look at people who joined EA in 2021 or later (the last 3 years seems decently recent to me, I don't have the sense that 80K's recruitment has changed so much in that time frame, n=1059), we see:

  • Global Poverty
    • 80,000 Hours
      • 35.8% rated Global Poverty 4/5
      • 34.2% rated Global Poverty 5/5
    • All respondents
      • 40.2% rated Global Poverty 4/5
      • 33.2% rated Global Poverty 4/5
  • Animal Welfare
    • 80,000 Hours
      • 26.8% rated Animal Welfare 4/5
      • 8.9% rated Animal Welfare 5/5
    • All respondents
      • 29.7% rated Animal Welfare 4/5
      • 13.7% rated Animal Welfare 5/5

Bob sees little reason to reconsider the trade-off, especially since ChatGPT seems to have vindicated 80,000 hours’ prior belief that AI was going to be a big deal


ChatGPT is just the tip of the iceberg here.

GPT4 is significantly more powerful than 3.5. Google now has a multi-modal model that can take in sound, images and video and a context window of up to a million tokens. Sora can generate amazing realistic videos. And everyone is waiting to see what GPT5 can do.

Further, the Center for AI Safety open letter has demonstrated that it isn't just our little community that is worried about these things, but a large number of AI experts.

Their 'AI is going to be a big thing' bet seems to have been a wise call, at least at the current point in time. Of course, I'm doing AI Safety movement building, so I'm a bit biased here, and maybe we'll think differently down the line, but right now they're clearly ahead.

I’m one of the people interviewed by Joel and I’d like to share a bit about my experience, as it could serve as an additional datapoint for how this report was constructed. 

For some context, I’m on the Groups Team at CEA, where I manage the Community Building Grants program. I reached out to Joel to see how I could support his research, as I was in favour of somebody looking into this question. 

The key points I’d like to bring up are: 

  1. Before going into the call, I was under the impression that it would be a chat, which bears for me the meaning of being more informal. But in practice the call felt more like answering a series of quite closed [and occasionally suggestive] questions. I did not make notes during the call, unfortunately, but this was the general vibe that I noted down after the call. 
  2. I didn’t realise that what I said in the call would get interpreted and shared as “CEA’s views.” Had I known that this would happen, I would have approached the call differently and caveated much more heavily that I could only speak for myself. 
  3. Relatedly, I wasn't informed that some information from the conversation would get shared publicly. While I know that in journalism it is a norm to not always clarify information, I had different expectations for a research report. I think that if I had known about being included in a public publication, I would have expected to get the chance to review how my comments were presented before this went live, which I did not get. I wish I'd had that opportunity to clarify potential misunderstandings and flag sensitive information.

I'm surprised by the scepticism re 80k. The OP EA/LT survey from 2020 seems to suggest one can be quite confident of the positive impact 80k had on the small (but important) population surveyed. As the authors noted in their summary:

Among the organizations, pieces of content, and meetup groups/discussion platforms we explicitly asked about, the ones that seemed to have been most impactful on our respondents robustly, across multiple metrics, were 80,000 Hours, CEA, local and university EA groups, FHI, Nick Bostrom’s works, Eliezer Yudkowsky’s works, and Peter Singer’s works.

Hey, Arden from 80,000 Hours here – 

I haven't read the full report, but given the time sensitivity with commenting on forum posts, I wanted to quickly provide some information relevant to some of the 80k mentions in the qualitative comments, which were flagged to me.

Regarding whether we have public measures of our impact & what they show

It is indeed hard to measure how much our programmes counterfactually help move talent to high impact causes in a way that increases global welfare, but we do try to do this.

From the 2022 report the relevant section is here. Copying it in as there are a bunch of links. 

We primarily use six sources of data to assess our impact:

  1. Open Philanthropy EA/LT survey
  2. EA Survey responses
  3. The 80,000 Hours user survey. A summary of the 2022 user survey is linked in the appendix. 
  4. Our in-depth case study analyses, which produce our top plan changes and DIPY estimates (last analysed in 2020). 
  5. Our own data about how users interact with our services (e.g. our historical metrics linked in the appendix). 
  6. Our and others' impressions of the quality of our visible output. 

Overall, we’d guess that 80,000 Hours continued to see diminishing returns to its impact per staff member per year. [But we continue to think it's still cost-effective, even as it grows.]

Some elaboration: 

  • DIPY estimates are our measure of contractual career plan shifts we think will be positive for the world. Unfortunately it's hard to get an accurate read on counterfactuals and response rates, so these are only very rough estimates & we don't put that much weight on them.
  • We report on things like engagement time & job board clicks as *lead metrics* because we think they tend to flow through to counterfactual high impact plan changes, & we're able to measure them much more readily.
  • Headlines from some of the links above: 
    • From our own survey (2138 respondents):
      • On the overall social impact that 80,000 Hours had on their career or career plans, 
        • 1021 (50%) said 80,000 Hours increased their impact
          • Within this we identified 266 who reported >30% chance of 80,000 Hours causing them to taking a new jobs or graduate course (a “criteria based plan change”)
        • 26 (1%) said 80,000 Hours reduced their impact.
          • Themes in answers were demoralisation and causing career choices that were a poor fit
    • Open Philanthropy's EA/LT survey was aimed at asking their respondents "​​“What was important in your journey towards longtermist priority work?” – it has a lot of different results and feels hard to summarise, but it showed a big chunk of people considered 80k a factor in ending up working where they are.
    • The 2020 EA survey link says "More than half (50.7%) of respondents cited 80,000 Hours as important for them getting involved in EA". (2022 says something similar

Regarding the extent to which we are cause neutral & whether we've been misleading about this

We do strive to be cause neutral, in the sense that we try to prioritize working on the issues where we think we can have the highest marginal impact (rather than committing to a particular cause for other reasons). 

For the past several years we've thought that the most pressing problem is AI safety, so have put much of our effort there (Some 80k programmes focus on it more than others – I reckon for some it's a majority, but it hasn't been true that as an org we “almost exclusively focus on AI risk.” (a bit more on that here.))

In other words, we're cause neutral, but not cause *agnostic* - we have a view about what's most pressing. (Of course we could be wrong or thinking about this badly, but I take that to be a different concern.)

The most prominent place we describe our problem prioritization is our problem profiles page – which is one of our most popular pages. We describe our list of issues this way: "These areas are ranked roughly by our guess at the expected impact of an additional person working on them, assuming your ability to contribute to solving each is similar (though there’s a lot of variation in the impact of work within each issue as well). (Here's also a past comment from me on a related issue.) 

Regarding the concern about us harming talented EAs by causing them to choose bad early career jobs

To the extent that this has happened this is quite serious – helping talented people have higher impact careers is our entire point! I think we will always sometimes fail to give good advice (given the diversity & complexity of people's situations & the world), but we do try to aggressively minimise negative impacts, and if people think any particular part of our advice is unhelpful, we'd like them to contact us about it! (I'm arden@80000hours.org & I can pass them on to the relevant people.)

We do also try to find evidence of negative impact, e.g. using our user survey, and it seems dramatically less common than the positive impact (see the stats above), though there are of course selection effects with that kind of method so one can't take that at face value!

Regarding our advice on working at AI companies and whether this increases AI risk

This is a good worry and we talk a lot about this internally! We wrote about this here.

The 2020 EA survey link says "More than half (50.7%) of respondents cited 80,000 Hours as important for them getting involved in EA". (2022 says something similar

 

I would also add these results, which I think are, if anything, even more relevant to assessing impact:

Hi Arden,

Thanks for engaging. 

(1) Impact measures: I'm very appreciative of the amount of thought that went into developing the DIPY measure. The main concern (from the outside) with respect to DIPY is that it is critically dependent on the impact-adjustment variable - it's probably the single biggest driver of uncertainty (since causes can vary by many magnitudes). Depending on whether you think the work is impactful (or if you're sceptical, e.g. because you're an AGI sceptic or because you're convinced of the importance of preventing AGI risk but worried about counterproductivity from getting people into AI etc), the estimate will fluctuate very heavily (and could be zero or significantly negative). From the perspective of an external funder, it's hard to be convinced of robust cost-effectiveness (or speaking for myself, as a researcher, it's hard to validate).

(2) I think we would both agree that AGI (and to a lesser extent, GCR more broadly) is 80,000 Hour's primary focus.

I suppose the disagreement then is the extent to which neartermist work gets any focus at all. This is to some extent subjective, and also dependent on hard-to-observe decision-making and resource-allocation done internally. With (a) the team not currently planning to focus on neartermist content for the website (the most visible thing), (b) the career advisory/1-1 work being very AGI-focused too (to my understanding), and (c) fundamentally, OP being 80,000 Hour's main funder, and all of OP's 80k grants being from the GCR capacity building team over the past 2-3 years - I think from an outside perspective, a reasonable assumption is that AGI/GCR is >=75% of marginal resources committed. I exclude the job board from analysis here because I understand it absorbs comparatively little internal FTE right now.

The other issue we seem to disagree on is whether 80k has made its prioritization sufficiently obvious. It appreciate that this is somewhat subjective, but it might be worth erring on the side of being too obvious here - I think the relevant metric would be "Does a average EA who looks at the job board or signs up for career consulting understand that 80,000 Hours prefers I prioritize AGI?", and I'm not sure that's the case right now.

(3) Bad career jobs - this was a concern aired, but we didn't have too much time to investigate it - we just flag it out as a potential risk for people to consider.

(4) Similarly, we deprioritized the issue of whether getting people into AI companies worsens AI risk. We leave it up to potential donors to be something they might have to weigh and consider the pros and cons of (e.g. per Ben's article) and to make their decisions accordingly.

Hey Joel, I'm wondering if you have recommendations on (1) or on the transparency/clarity element of (2)?

(Context being that I think 80k do a good job on these things, and I expect I'm doing a less good job on the equivalents in my own talent search org. Having a sense of what an 'even better' version might look like could help shift my sort of internal/personal overton window of possibilities.)

Hi Jamie,

For (1) I'm agree with 80k's approach in theory - it's just that cost-effectiveness is likely heavily driven by the cause-level impact adjustment - so you'll want to model that in a lot of detail.

For (2), I think just declaring up front what you think is the most impactful cause(s) and what you're focusing on is pretty valuable? And I suppose when people do apply/email, it's worth making that sort of caveat as well. For our own GHD grantmaking, we do try to declare on our front page that our current focus is NCD policy and also if someone approaches us raising the issue of grants, we make clear what our current grant cycle is focused on.

Hope my two cents is somewhat useful!

Makes sense on (1). I agree that this kind of methodology is not very externally legible and depends heavily on cause prioritisation, sub-cause prioritisation, your view on the most impactful interventions, etc. I think it's worth tracking for internal decision-making even if external stakeholders might not agree with all the ratings and decisions. (The system I came up with for Animal Advocacy Careers' impact evaluation suffered similar issues within animal advocacy.)

For (2), I'm not sure why you don't think 80 do this. E.g. the page on "What are the most pressing world problems?" has the following opening paragraph:

We aim to list issues where each additional person can have the most positive impact. So we focus on problems that others neglect, which are solvable, and which are unusually big in scale, often because they could affect many future generations — such as existential risks. This makes our list different from those you might find elsewhere.

Then the actual ranking is very clear: AI 1, pandemics 2, nuclear war 3, etc. 

And the advising page says quite prominently "We’re most helpful for people who... Are interested in the problems we think are most pressing, which you can read about in our problem profiles." The FAQ on "What are you looking for in the application?" mentions that one criterion is "Are interested in working on our pressing problems".

Of course it would be possible to make it more prominent, but it seems like they've put these things pretty clearly on the front. 

It seems pretty reasonable to me that 80k would want to talk to people who seem promising but don't share all the same cause prio views as them; supporting people to think through cause prio seems like a big way they can add value. So I wouldn't expect them to try to actively deter people who sign up and seem worth advising but, despite the clear labelling on the advising page, don't already share the same cause prio rankings as 80k. You also suggest "when people do apply/email, it's worth making that sort of caveat as well", and that seems in the active deterrence ballpark to me; to the effect of 'hey are you sure you want this call?'

On (2). If you go to 80k's front page (https://80000hours.org/), there is no mention that the organizational's focus is AGI or that they believe it to be the most important cause. For the other high-level pages accessible from the navigation bar, things are similar not obvious. For example, in "Start Here", you have to read 22 paragraphs down to understand 80k's explicit prioritization of x-risk over other causes. In the "Career Guide", it's about halfway down the page. If the 1-1 advising tab, you have to go down to the FAQs at the bottom of the page, and even then it only refers to "pressing problems" and links back to the research page. And on the research page itself, the issue is that it doesn't give a sense that the organization strongly recommends AI over the rest, or that x-risk gets the lion's share of organizational resources.

I'm not trying to be nitpicky, but trying to convey that a lot of less engaged EAs (or people who are just considering impactful careers) are coming in, reading the website, and maybe browsing the job board or thinking of applying for advising - without realizing just how convinced on AGI 80k is (and correspondingly, not realizing how strongly they will be sold on AGI in advisory calls). This may not just be less engaged EAs too, depending on how you defined engaged - like I was reading Singer since two decades ago; have been a GWWC pledger since 2014; and whenever giving to GiveWell have actually taken the time to examine their CEAs and research reports. And yet until I actually moved into direct EA work via the CE incubation program, I didn't realize how AGI-focused 80k was.

People will never get the same mistaken impression when looking at Non-Linear or Lightcone or BERI or SFF. I think part of the problem is (a) putting up a lot of causes on the problems page, which gives the reader the impression of a big tent/broad focus, and (b) having normie aesthetics (compare: longtermist websites). While I do think it's correct and valuable to do both, the downside is that without more explicit clarification (e.g. what Non-Linear does, just bluntly saying on the front page in font 40: "We incubate AI x-risk nonprofits ​by connecting founders with ideas, funding, and mentorship"), the casual reader of the website doesn't understand that 80k basically works on AGI.

Yeah many of those things seem right to me.

I suspect the crux might be that I don't necessarily think it's a bad thing if "the casual reader of the website doesn't understand that 80k basically works on AGI". E.g. if 80k adds value to someone as they go through the career guide, even if they don' realise that "the organization strongly recommends AI over the rest, or that x-risk gets the lion's share of organizational resources", is there a problem?

I would be concerned if 80k was not adding value. E.g. I can imagine more salesly tactics that look like making a big song and dance about how much the reader needs their advice, without providing any actual guidance until they deliver the final pitch, where the reader is basically given the choice of signing up for 80k's view/service, or looking for some alternative provider/resource that can help them. But I don't think that that's happening here.

I can also imagine being concerned if the service was not transparent until you were actually on the call, and then you received some sort of unsolicited cause prioritisation pitch. But again, I don't think that's what's happening; as discussed, it's pretty transparent on the advising page and cause prio page what they're doing.

Very interesting report. It provided a lot of visibility into how these funders think.

Geographically, India may be a standout opportunity for getting talent to do research/direct work in a counterfactually cheap way.

I would have embraced that more in the past, but I'm a lot more skeptical of this these days and I think that if it were possible then it would have worked out. For many tasks, EA wants the best talent that is available. Top talent is best able to access overseas opportunities and so the price is largely independent of current location.

In terms of age prioritization, it is suboptimal that EA focuses more on
outreach to university students or young professionals as opposed to mid-career
people with greater expertise and experience.

I agree that there is a lot of alpha in reaching out to mid-career professionals - if you are able to successfully achieve this. This work is a lot more challenging - mid-career professionals are often harder to reach, less able to pivot and have less time available to upskill. Less people are able to do this kind of outreach because these professionals may take current students or recent grads less seriously. So for a lot of potential movement builders, focusing on students or young professionals is a solid play because it is the best combination of impact and personal fit.

The report writes "80,000 Hours in particular, has been singled out for its work not having a clear positive impact, despite the enormous sums they have spent.

As an on-the-ground community builder, I'm skeptical of your take. So many people that I've talked to became interested in EA or AI safety through 80,000 Hours. Regarding: "misleading the community that it is cause neutral while being almost exclusively focused on AI risk", I was more concerned about this in the past, but I feel that they're handling this pretty well these days. Would be quite interested to hear if you had specific ways you thought they should change. Regarding, causing early career EA's to choose suboptimal early career jobs that mess up their CVs, I'd love to hear more detail on this if you can. Has anyone written up a post on this?

On funding Rethink Priorities specifically – a view floated is that there is value in RP as a check on OP, and yet if OP doesn't think RP is worth funding beyond a certain point, it's hard to gainsay them

Rethink Priorities seems to be a fairly big org - especially taking into account that it operates on the meta-level - so I understand why OpenPhil might be reluctant to spend even more money there. I suspect that there's a feeling that even if they are doing good work, that meta work should only be a certain portion of their budge. I wouldn't take that as a strong signal.

In practice, it's unclear if the community building actually translates to money/talent moved, as opposed to getting people into EA socially.

As someone on the ground, I can say that community building has translated to talent moved at the very least. Money moved is less visible to me because a lot of people aren't shouting about the pledges from the rooftops, plus the fact that donations are very heavy-tailed. Would love to hear some thoughtful discussion of how this could be better measured.

What would interest OP might be a project about getting in more people who are doing the best things in GHW (e.g. Buddy Shah, Eirik Mofoss).

Very curious about what impressed Open Philanthropy about these people since I'm not familiar with their work. Would be keen to learn more though!

Hi Chris,

Just to respond to the points you raised

(1) With respect to prioritize India/developing country talent, it probably depends on the type of work (e.g. direct work in GHD/AW suffers less for this), but in any case, the pool of talent is big, and the cost savings are substantial, so it might be reasonably to go this route regardless.

(2) Agreed that it's challenging, but I guess it's a chicken vs egg problem - we probably have to start somewhere (e.g. HIP etc does good work in the space, we understand).

(3) For 80k, see my discussion with Arden above - AGB's views are also reasonably close to my own.

(4) On Rethink - to be fair, our next statement after that sentence is "This objection holds less water if one is disinclined to accept OP's judgement as final." I think OP's moral weights work, especially, is very valuable.

(5) There's a huge challenge over valuing talent, especially early career talent (especially if you consider the counterfactual being earning to give at a normal job). One useful heuristic is: Would the typical EA organization prefer an additional 5k in donations (from an early career EA giving 10% of their income annually) or 10 additional job applications to a role? My sense from talking to organizations in the space is that (a) the smaller orgs are far more funding constrained, so prefer the former, and (b) the bigger orgs are more agnostic, because funding is less a challenge but also there is a lot of demand for their jobs anyway.

(6) I can't speak for OP specifically, but I (and others in the GHD policy space I've spoken to) think that Eirik is great. And generally, in GHD, the highest impact work is convincing governments to change the way they do things, and you can't really do that without positions of influence.

For 1: A lot of global health and development is much less talent-hungry than animal welfare work or x-risk work. Take for example the Against Malaria Foundation. They receive hundreds of millions of dollars, but they only have a core team of 13. Sure you need a bunch of people to hand out bed nets, but the requirements for that aren't that tight and sure you need some managers, but lots of people are capable of handling this kind of logistics, you don't really have to headhunt them. I suppose this could change if there was more of a pivot into policy where talent really matters. However, in that case, you would probably want people from the country whose policy you want to influence, moreso then thinking about cost.

For 5: It's not clear to me that the way you're thinking about this makes sense to me. If you're asking about the trade-off between direct work and donations, it seems as though we should ask about $5k from the job vs. a new candidate who is better than your current candidate as, in a lot of circumstances, they will have the option of doing earn to give so long as they don't take an EA job (I suppose there is the additional factor of how much chasing EA jobs detracts from chasing earn to give jobs).

1) It agree that policy talent is important but comparatively scarce, even in GHD. It's the biggest bottleneck that Charity Entrepreneurship is facing on incubating GHD policy organizations right now, unfortunately.

5) I don't think it's safe to assume that the new candidate is better than your current candidate? While I agree that's fine for dedicated talent pipeline programmes, I'm not confident of making this assumption for general community building, is by its nature less targeted and typically more university/early-career oriented.

My point was that presumably the org thinks they're better if they decide to hire them as opposed to the next best person.

I apologize if we're talking at cross purposes, but the original idea I was trying to get across is that when valuing additional talent from community building, there is the opportunity cost of a non-EA career where you just give. So basically you're comparing (a) the value of money from that earning to give vs (b) the value of the same individual trying for various EA jobs.

The complication is that (i) the uncertainty of the individual really following through on the intention to earn to give (or going into an impactful career) applies to both branches; however, (ii) the uncertainty of success only applies to (b). If they really try to earn to give they can trivially succeed (e.g. give 10% of the average American salary - so maybe $5k, ignoring adjustments for lower salaries for younger individual and higher salaries for typically elite educated EAs). However, if they apply to a bunch of EA jobs, the aren't necessarily going to succeed (i.e. they aren't necessary going to be better than the counterfactual hire). So ultimately we're comparing the value an additional $5k annual donation vs additional ~10 applications of average quality to various organizations (depends on how many organizations an application will apply to per annum - very uncertain).

I also can't speak with certainty as to how organizations will choose, but my sense is that (a) smaller EA organizations are funding constrained and would prefer getting the money; while (b) larger EA organizations are more agnostic because they have both more money and the privilege of getting the pick of the crop for talent (c.f. high demand for GiveWell/OP jobs).

Okay, I guess parts of that framework make a bit more sense now that you've explained it.

At the same time, it feels that people can always decide to earn to give if they fail to land an EA-relevant gig, so I'm not sure why you're modeling it as a $5k annual donation vs. a one-time $5k donation for someone spending a year focusing on upskilling for EA roles. Maybe you could add an extra factor for the slowdown in their career advancement, but $50k extra per year is unrealistic.

I think it's also worth considering that there are selection effects here. So insofar as EA promotes direct work, people with higher odds of being successful in landing a direct work position are more likely to pursue that and people with better earn-to-give potential are less likely to take the advice.

Additionally, I wonder whether the orgs you surveyed understood ten additional applications as ten additional average applications or ten additional applications from EA's (more educated and valued-aligned than the general population) who were dedicated enough to actually follow through on earning to give. 

I think you're right in pointing out the limitations of the toy model, and I strongly agree that the trade-off is not as stark as it seems - it's more realistic that we model it aa a delay from applying to EA jobs before settling for a non EA job (and that this wont be like a year or anything)

However, I do worry that the focus on direct work means people generally neglect donations as a path to impact and so the practical impact of deciding to go for an EA career is that people decide not to give. An unpleasant surprise I got from talking to HIP and others in the space is that the majority of EAs probably don't actually give. Maybe it's the EA boomer in me speaking, but it's a fairly different culture compared to 10+ years ago where being EA meant you bought into the drowning child arguments and gave 10% or more to whatever cause you thought most important

I’m confused by the use of the term “expert” throughout this report. What exactly is the expertise that these individual donors and staff members are meant to be contributing? Something more neutral like ‘stakeholder’ seems more accurate.

Generally, they have a combination of the following characteristics: (a) a direct understanding of what their own grantmaking organization is doing and why, (b) deep knowledge of the object-level issue (e.g. what GHD/animal welfare/longtermist projects to fund, and (c) extensive knowledge of the overall meta landscape (e.g. in terms of what other important people/organizations there are, the background history of EA funding up to a decade in the past, etc).

I'd suggest using a different term or explicitly outlining how you use "expert" (ideally both in the post and in the report, where you first use the term) since I'm guessing that many readers will expect that if someone is called "expert" in this context, they're probably "experts in EA meta funding" specifically — e.g. someone who's been involved in the meta EA funding space for a long time, or someone with deep knowledge of grantmaking approaches at multiple organizations. (As an intuition pump and personal datapoint, I wouldn't expect "experts" in the context of a report on how to run good EA conference sessions to include me, despite the fact that I've been a speaker at EA Global a few times.) Given your description of "experts" above, which seems like it could include (for instance) someone who's worked at a specific organization and maybe fundraised for it, my sense is that the default expectation of what "expert" means in the report would this be mistaken. 


Relatedly, I'd appreciate it if you listed numbers (and possibly other specific info) in places like this: 

We interviewed numerous experts, including but not limited to staff employed by (or donors associated with) the following organizations: OP, EA Funds, MCF, GiveWell, ACE, SFF, FP, GWWC, CE, HLI and CEA. We also surveyed the EA community at large.

E.g. the excerpt above might turn into something like the following: 

We interviewed [10?] [experts], including staff at [these organizations] and donors who have supported [these organizations]. We also ran an "EA Meta Funding Survey" of people involved in the EA community and got 25 responses.

This probably also applies in places where you say things like "some experts" or that something is "generally agreed". (In case it helps, a post I love has a section on how to be (epistemically) legible.)

I’m inferring from other comments that AGB as an individual EtG donor is the “expert [who] funded [80k] at an early stage of their existence, but has not funded them since” mentioned in the report. If this is the case, how do individual EtG donors relate to the criteria you mention here?

Executive summary: The EA meta funding landscape saw rapid growth and contraction from 2021-2023, with Open Philanthropy being the largest funder and longtermism receiving the most funding, while the landscape faces high uncertainty in the short-to-medium term.

Key points:

  1. EA meta funding grew from $109M in 2021 to $193M in 2022, before shrinking to $117M in 2023, driven largely by changes in Open Philanthropy's spending.
  2. Funding allocation by cause, in descending order: longtermism ($274M) >> global health and development ($67M) > cross-cause ($53M) > animal welfare ($25M).
  3. Funding allocation by intervention, in descending order: other/miscellaneous ($193M) > talent ($121M) > prioritization ($92M) >> effective giving ($13M).
  4. The EA meta grantmaking landscape is highly concentrated, with core funders (Open Philanthropy, EA Funds, and SFF) making up 82% of total funding.
  5. Crucial considerations for prioritizing direct vs meta work include the relative cost-effectiveness of the direct work and whether opportunities are getting better or worse.
  6. The meta funding landscape faces high uncertainty in the short-to-medium term due to turnover at Open Philanthropy, uncertainty at EA Funds, and the ongoing consequences of the FTX collapse.

 

 

This comment was auto-generated by the EA Forum Team. Feel free to point out issues with this summary by replying to the comment, and contact us if you have feedback.

[with respect to talent] In terms of age prioritization, it is suboptimal that EA focuses more on outreach to university students or young professionals as opposed to mid-career people with greater expertise and experience.

 

I think this is conditional on the object-level positions available being reasonably well-suited to mid-career folks. For instance, job security becomes increasingly important as one ages. Even to the extent that successful mid-career folks face a risk of being let go in their current positions, they know they should be able to find another position in their field pretty easily. Likewise, positions in some cause areas may not pay well enough to attract much interest from successful mid-career professionals in certain geographic locations who have a kid and a mortgage. Those areas are probably better off with the current recruiting focus, at least in developed countries.

I think this is a legitimate concern, but it's not clear to me that it outweighs the benefits, especially for roles where experience is essential.

Curated and popular this week
Relevant opportunities