This is a special post for quick takes by NickLaing. Only they can create top-level comments. Comments here also appear on the Quick Takes page and All Posts page.
Sorted by Click to highlight new quick takes since:

Im intrigued where people stand on the threshold where farmed animal lives might become net positive? I'm going to share a few scenarios i'm very unsure about and id love to hear thoughts or be pointed towards research on this.

  1. Animals kept in homesteads in rural Uganda where I live. Often they stay inside with the family at night, then are let out during the day to roam free along the farm or community. The animals seem pretty darn happy most of the time for what it's worth, playing and galavanting around. Downsides here include poor veterinary care so sometimes parasites and sickness are pretty bad and often pretty rough transport and slaughter methods (my intuition net positive).

  2. Grass fed sheep in New Zealand, my birth country. They get good medical care, are well fed on grass and usually have large roaming areas (intuition net positive)

  3. Grass fed dairy cows in New Zealand. They roam fairly freely and will have very good vet care, but have they calves taken away at birth, have constantly uncomfortably swollen udders and are milked at least twice daily. (Intuition very unsure)

  4. Free range pigs. Similar to the above except often space is smaller but they do get little houses. Pigs are far more intelligent than cows or sheep and might have more intellectual needs not getting met. (Intuition uncertain)

Obviously these kind of cases make up a small proportion of farmed animals worldwide, with the predominant situation - factory farmed animals likely having net negative lives.

I know that animals having net positive lives far from justifies farming animals on it's own, but it seems important for my own decision making and for standing on solid ground while talking with others about animal suffering.

Thanks for your input.

It's really hard to judge whether a life is net positive. I'm not even sure when my own life is net positive—sometimes if I'm going through a difficult moment, as a mental exercise I ask myself, "if the rest of my life felt exactly like this, would I want to keep living?" And it's genuinely pretty hard to tell. Sometimes it's obvious, like right at this moment my life is definitely net positive, but when I'm feeling bad, it's hard to say where the threshold is. If I can't even identify the threshold for myself, I doubt I can identify it in farm animals.

If I had to guess, I'd say the threshold is something like

  • if the animals spend most of their time outdoors, their lives are net positive
  • if they spend most of their time indoors (in crowded factory farm conditions, even if "free range"), their lives are net negative

it seems important for my own decision making and for standing on solid ground while talking with others about animal suffering.

To this point, I think the most important things are

  1. whatever the threshold is, factory-farmed animals clearly don't meet it
  2. 99% of animals people eat are factory-farmed (in spite of people's insistence that they only eat meat from their uncle's farm where all of the animals are treated like their own children etc)

That's really interesting on your own life. Even in the midst of my worst emotional states (emotional not physical pain,) I would still feelI'm on the positive side of the ledger.

Yes I agree on your list 2 points, those are the most important in general

In Northern Uganda here though. the majority of animals people eat (not often, many people eat meat once or twice a month) have lives from my first example. In New Zealand almost all beef and sheep meat is from something like options 2 and 3, so I think the question has some relevance to a decent number of people.

Thanks for this. My view is the same as yours. The first two strike me as "net positive." I'm also unsure about what pigs and dairy cows need. I wouldn't be hugely surprised if they have either "net positive" or "net negative" lives, but I think it's most likely (80%+ chance) they are "net positive."

(Qualifying discussion of net value of existence with " " because I find such valuations always so fraught with uncertainty and I feel I owe other beings tremendous humility in this!) 

Just adding: the discussion of dairy cows, here and elsewhere, tends to focus on the experience of the adult cattle & the suffering for them of being milked, deprived of their babies, etc.

But it's not implausible to me that the majority of the disvalue from dairy is in the lives of the calves born to dairy cows. In typical milk-producing operations, adult cows have 1 calf every 18 months or so; 50% of them are male, and so are killed within a few hours to a few months after birth.

(& these lives more likely to be net negative because they have less time to experience positive things to outweigh the terror and pain of death. Undoubtedly, some of their deaths will be quite quick, but others are slow and brutal.)

(Also, veal calves are treated very badly - intense confinement to reduce movement to keep the meat tender, dietary restriction to keep the meat pale, individual confinement in a tiny 'hutch', etc.)

(& let's not forget the fetal calves who are still gestating when their mothers go to slaughter. They're killed slowly, if they ever get purposefully slaughtered at all rather than just left to asphyxiate. Obviously, it's unclear whether they're conscious, but I've read accounts of them moving, opening eyes, trying to breathe, etc.).

Thanks Bella, this has crossed my mind and definitely updates me towards dairy farmed cows in New Zealand being more likely to be net negative. I'm not sure whether the veal thing happens in New Zealand though I'll look into it.

huw
5
1
0
1

I’m always surprised to see sheep get lumped in with cows in discussions of farmed animal welfare (ex. the SSC Adversarial Collaboration). Sure, it’s not a terrible proxy, but sheep are often freer, need to be regularly shorn to avoid overheating, and usually die of natural causes. There are definitely some practices which are awful, but sheep are quite hard to optimise in the same way we’ve done with pigs & chickens, or even cows.

However, we eat them when they’re babies so maybe it swings in the absolute other direction.

Nice point Huw, I agree. I wasn't trying to lump them together exactly, I agree its very different. Does eating them when they are babies necessarily mean a net negative life though, if the slaughter is humane? It does seem like a weird question though...

The idea behind why eating babies is more likely to be net negative is that there's a shorter lifespan of positive experiences to balance out the terror and pain of death.

From my experience watching lots of slaughterhouse footage and reading accounts from workers, even the best humane conditions still involve, routinely, a (shorter or longer) period in which the animal goes through the process of dying. This is probably pretty bad. If they only lived for a few weeks before that, it's harder to imagine it's a good deal overall.

Under some frameworks, you’d be depriving them of many years of happy life; but then again, if you didn’t kill them as children they probably would never have been born for food. Here we’d be getting too deep into the moral philosophy for me to have a confident take 😅. Interesting nonetheless.

"but it seems important for my own decision making and for standing on solid ground while talking with others about animal suffering."

I'm highly skeptical of this - why do you think it is important for your own moral decision making? It seems to me that whether farmed animals lives are worth living or not is irrelevant - either way we should try to improve their conditions, and the best ways of doing that seem to be: a boycott & political pressure (I would argue that the two work well together).

By analogy, no one raises the question of whether the lives of people living in extreme poverty, or working in sweatshops and so on, are worth living, because it's simply irrelevant.

This seems relevant to any intervention premised on "it's good to reduce the amount of net-negative lives lived."

If factory-farmed chickens have lives that aren't worth living, then one might support an intervention that reduces the number of factory-farmed chickens, even if it doesn't improve the lives of any chickens that do come to exist. (It seems to me this would be the primary effect of boycotts, for instance, although I don't know empirically how true that is.)

I agree that this is irrelevant to interventions that just seek to improve conditions for animals, rather than changing the number of animals that exist. Those seem equally good regardless of where the zero point is.

I suppose I agree with this. And I've been mulling over why it still seems like the wrong way to think about it to me, and I think it's that I find it rather short-termist. In the short term if farms shut down they might be replaced with nature, with even less happy animals, it's true. But in the long term opposing speciesism is the only way to achieve a world with happy beings. Clearly the kinds of farms @NickLaing is talking about, with lives worth living but still pretty miserable, are not optimal. Figuring out whether they are worth living or not seems only relevant to trying to reduce suffering in the short term, but not so much in the long term, because in the long term this isn't what we want anyway.

I'm a little confused as to why we consider the leaders of AI companies (Altman, Hassabis, Amodei etc.) to be "thought leaders" in the field of AI safety in particular. Their job descriptions are to grow the company and increase shareholder value, so their public persona and statements has to reflect that. Surely they are far too compromised for their opinions to be taken too seriously, they couldn't make strong statements against AI growth and development even if they wanted to, because of their job and position.

The recent post "Sam Altman's chip ambitions undercut OpenAI's safety strategy"  seems correct and important, while also almost absurdly obvious - the guy is trying to grow his company and they need more and better chips. We don't seriously listen to big tobacco CEOs about the dangers of smoking, or Oil CEOs about the dangers of climate change, or Factory Farming CEOs about animal suffering, so why do we seem to take the opinions of AI bosses about safety even in moderate good faith? The past is often the best predictor of the future, and the past here says that CEOs will grow their companies, while trying however possible to maintain public goodwill as to minimise the backlash.

I agree that these CEOs could be considered thought leaders in AI in general and the Future and potential of AI, and their statements about safety and the future are critically practically important and should be engaged with seriously. But I don't really see the point of engaging with them as thought leaders in the AI safety discussion, it would make more sense to me to rather engage with intellectuals and commentators who can fully and transparently share their views without crippling levels of compromisation.

I'm interested to hear though arguments in favour of taking their thoughts more seriously.

Who is considering Altman and Hassabis thought leaders in AI safety? I wouldn't even consider Altman a thought leader in AI - his extraordinary skill seems mostly social and organizational. There's maybe an argument for Amodei, as Anthropic is currently the only one of the company whose commitment to safety over scaling is at least reasonably plausible.

Thanks that's a helpful perspective and I would be happy if it was true that they weren't considered AI safety thought leaders. I do feel like they are often seen this way though in the public sphere, and sometimes here on the forum too.

I realize that my question sounded rethorical, but I'm actually interested in your sources or reasons for your impression. I certainly don't have a good idea of the general opinion and the media I consume is biased towards what I consider reasonable takes. That being said, I haven't encountered the position you're concerned about very much and would be interested to be hear where you did. Regarding this forum, I imagine one could read into some answers, but overall I don't get the impression that the AI CEO's are seen as big safety proponents.

I think thoughtleader sometimes means "has thoughts at the leading edge" and sometimes mean "leads the thoughts of the herd on a subject" and that there is sometimes a deliberate ambiguity between the two.

The Happier Lives Institute have helped many people (including me) open their eyes to Subjective Wellbeing and perhaps even update us to the potential value of SWB. The recent heavy discussion (60+ comments) on their fundraising thread disheartened me. Although I agree with much of the criticism against them, the hammering they took felt at best rough and perhaps even unfair. I'm not sure exactly why I felt this way, but here are a few ideas.

  • (High certainty) HLI have openly published their research and ideas, posted almost everything on the forum and engaged deeply with criticism which is amazing - more than perhaps any other org I have seen. This may (uncertain) have hurt them more than it has helped them.
  • (High certainty) When other orgs are criticised or asked questions, they often don't reply at all, or get surprisingly little criticism for what I and many EAs might consider poor epistemics and defensiveness in their posts (for charity I'm not going to link to the handful I can think of). Why does HLI get such a hard time while others get a pass? Especially when HLI's funding is less than many of orgs that have not been scrutinised as much.
  • (Low certainty) The degree of scrutiny and analysis of some development orgs in general like HLI seems to exceed that of AI orgs, Funding orgs and Community building orgs. This scrutiny has been intense- more than one amazing statistician has picked apart their analysis. This expert-level scrutiny is fantastic, I just wish it could be applied to other orgs as well. Very few EA orgs (at least that have been posted on the forum) produce full papers with publishable level deep statistical analysis like HLI have at least attempted to do. Does there need to be a "scrutiny rebalancing" of sorts. I would rather other orgs got more scrutiny, rather than development orgs getting less.

Other orgs might see threads like the HLI funding thread hammering and compare it with other threads where orgs are criticised and don't engage, so the thread falls off the frontpage. Orgs might reasonably decide that high degrees of transparency and engagement might do them net harm rather than good. This might not be good for anyone

Do you agree/disagree? And what could we do to make the situation better?

I think it's fairest to compare HLI's charity analysis with other charity evaluators like Givewell, ACE, and Giving Green.

Giving Green has been criticised regularly and robustly (just look up any of their posts). Givewell publish their analysis and engage with criticism; HLI themselves have actually criticised them pretty robustly! I don't know about ACE because I don't stay up to date on animals but I bet it's similar there.

The dynamics are quite different for example in charitable foundations where they don't need to convince anyone to donate differently, or charities that deliver a service who only need to convince their funders to continue donating.

Thanks Kristen for this clear and concise reply. This comparison with the experience of other charity evaluators has shifted my opinion on this somewhat nice one.

It seems a bit of a pity that they should receive significantly more scrutiny than charities or foundations though. In an ideal world everyone should be transparent and heavily scrutinised but it does make sense that the incentives might not be there for other orgs...

I agree that more orgs should get this kind of scrutiny. I agree that we are likely to blindly trust orgs that don't transparently discuss their inst workings, which is super sad.

Interesting reflection on Mental health providers too, be that's not a world I know!

This argument I struggle with...

"I don't think the problem is that HLI got too much hate for fucking up, it's that everyone else gets too little hate for being opaque"

I realize you are probably beinga bit tongue in cheek, but I think we could criticise and discuss while being more encouraging and positive. We are all human, to and I'm not sure piling on the "hate" will necessarily lead to improvement in epistemics and rigorous analysis.

[This comment is no longer endorsed by its author]Reply

I don't think this was a reply to me?

Sorry accidentally got confused and sister a comment my bad!

I don't know about ACE because I don't stay up to date on animals but I bet it's similar there.

I am a bit more familiar with ACE, and my impression is that you are right.

HLI fucked up their analysis, but because it was public we found out about it. Most EAs are too fearful to expose their work to scrutiny. Compare them to others who work on mental health within EA...

Most coaches and therapists in EA don't do any rigorous testing of whether what they are doing actually works. They don't even allow you to leave public reviews for them. I think we're the only organisation to even have a TrustPilot!!!

I don't think the problem is that HLI got too much hate for fucking up, it's that everyone else gets too little hate for being opaque.

Now HLI have been dragged through the mud, you can bet your ass they won't be making the same mistakes again. So long as they keep being transparent, they'll keep learning and growing as an org. Others will keep making the same mistakes indefinitely, only we'll never know about it and will continue blindly trusting them.

I agree that more orgs should get this kind of scrutiny. I agree that we are likely to blindly trust orgs that don't transparently discuss their inst workings, which is super sad.

Interesting reflection on Mental health providers too, be that's not a world I know!

This argument I struggle with...

"I don't think the problem is that HLI got too much hate for fucking up, it's that everyone else gets too little hate for being opaque"

I realize you are probably beinga bit tongue in cheek, but I think we could criticise and discuss while being more encouraging and positive. We are all human, to and I'm not sure piling on the "hate" will necessarily lead to improvement in epistemics and rigorous analysis.

Although I agree with much of the criticism against them, the hammering they took felt at best rough and perhaps even unfair.

One general problem with online discourse is that even if each individual makes a fair critique, the net effect of a lot of people doing this can be disproportionate, since there's a coordination problem. That said, a few things make me think the level of criticism leveled at HLI was reasonable, namely:

  1. HLI was asking for a lot of money ($200k-$1 million).
  2. The critiques people were making seemed (generally) unique, specific, and fair.
  3. The critiques came after some initial positive responses to the post, including responses to the effect of "I'm persuaded by this; how can I donate?"

Does there need to be a "scrutiny rebalancing" of sorts. I would rather other orgs got more scrutiny, rather than development orgs getting less.

I agree with you that GHD organizations tend to be scrutinized more closely, in large part because there is more data to scrutinize. But there is also some logic to balancing scrutiny levels within cause areas. When HLI solicits donations via Forum post, it seems reasonable to assume that donations they receive more likely come out of GiveWell's coffers than MIRI's. This seems like an argument for holding HLI to the GiveWell standard of scrutiny, rather than the MIRI standard (at least in this case).

That said, I do think it would be good to apply stricter standards of scrutiny to other EA organizations, without those organizations explicitly opening themselves up to evaluation by posting on the Forum. I wonder if there might be some way to incentivize this kind of review.

When HLI solicits donations via Forum post, it seems reasonable to assume that donations they receive more likely come out of GiveWell's coffers than MIRI's. This seems like an argument for holding HLI to the GiveWell standard of scrutiny, rather than the MIRI standard (at least in this case).

I am concerned that rationale would unduly entrench established players and stifle innovation. Young orgs on a shoestring budget aren't going to be able to withstand 2023 GiveWell-level scrutiny . . . and neither could GiveWell at the young-org stage of development.

Yeah, I should've probably been more precise: the criticism of HLI has mainly been leveled against their evaluation of a single organization's single intervention, whereas GW has evaluated 100+ programs, so my gut instinct is that it's fair to hold HLI's StrongMinds evaluation to the same ballpark level of scrutiny we'd hold a single GW evaluation to (and deworming certainly has been held to that standard). It might be unfair to expect an HLI evaluation to be at the level as a GW evaluation per dollar invested/hour spent (given that there's a learning curve associated with doing such evaluations and there's value associated with having multiple organizations do them), but this seems like—if anything—an argument for scrutinizing HLI's work more closely, since HLI is trying to climb a learning curve, and feedback facilitates this.

Which GiveWell evaluation(s) though? The ones on that spreadsheet range from the evaluations used to justify Top Charity status to decisions to deprioritize a potential program after a shallow review. Two deworming charities were until recently GiveWell Top Charities, and I believe Open Phil still makes significant grants to them (presumably in reliance on GiveWell's work).

In this post, HLI explicitly compares its evaluation of StrongMinds to GiveWell's evaluation of AMF, and says:

 "At one end, AMF is 1.3x better than StrongMinds. At the other, StrongMinds is 12x better than AMF. Ultimately, AMF is less cost-effective than StrongMinds under almost all assumptions. 

Our general recommendation to donors is StrongMinds."

This seems like an argument for scrutinizing HLI's evaluation of StrongMinds just as closely as we'd scrutinize GiveWell's evaluation of AMF (i.e., closely). I apologize for the trite analogy, but: if every year Bob's blueberry pie wins the prize for best pie at the state fair, and this year Jim, a newcomer, is claiming that his blueberry pie is better than Bob's, this isn't an argument for employing a more lax standard of judging for Jim's pie. Nor do I see how concluding that Jim's pie isn't the best pie this year—but here's a lot of feedback on how Jim can improve his pie for next year—undermines Jim's ability to win pie competitions going forward.

This isn't to say that we should expect the claims in HLI's evaluation to be backed by the same level of evidence as GiveWell's, but we should be able to take a hard look at HLI's report and determine that the strong claims made on its basis are (somewhat) justified.

Yes, agree that the language re: AMF justifies a higher level of scrutiny than would be warranted in its absence. Also, the AMF-related claim makes more moderate changes in the CEA bottom-line material than if the claims had been limited to stuff like: SM is more cost-effective than other predominately life-enhancing charities like GiveDirectly.

I think another factor is that HLI's analysis is not just below the level of Givewell, but below  a more basic standard. If HLI had performed at this basic standard, but below Givewell, I think strong criticism would have been unreasonable, as they are still a young and small org with plenty of room to grow. But as it stands the deficiencies are substantial, and a major rethink doesn't appear to be forthcoming, despite being warranted.

Probably a stupid question (probably just missed), can someone point me to where Givewell do a meta-analysis or similar depth of analysis as this HLI one. I can't seem to find it and I would be keen to do a quick compare myself.

I'm not aware of a GW analysis quite like this one, although I didn't go back and look at all its prior work.

In a situation like this, where GiveWell was considering StrongMinds as a top charity recommendation, it's almost certain that it would have first funded a bespoke RCT designed to address key questions for which the available literature was mixed or inconclusive. HLI doesn't have that luxury, of course. Moreover, what HLI is trying to measure is significantly harder to tease out than "how well do bednets work at saving lives" and similar questions.

I think those are relevant considerations that make comparing HLI's work to the "GiveWell standard" inappropriate. However, to acknowledge Ben's point, HLI's critics are alleging that the stuff that was missed was pretty obvious and that HLI hasn't responded appropriately when the missed stuff was pointed out. I lack the technical background and expertise to fully evaluate those claims.

My read is it wasn't the statistics they got hammered on misrepresenting other people's views of them as endorsements e.g. James Snowden's views. I will also say the AI side does get this criticism but not on cost-effectiveness but on things like culture war (AI Ethics vs. AI Safety) and dooming about techniques (e.g. working in a big company vs. more EA aligned research group and RLHF discourse). 

Thanks for the AI perspective!

Yes in that post the misrepresentation was part of the criticism they receive (which they engaged with and was at least partially corrected which is impressive) but I think the statistical analysis bore the most heavy overall criticism in that post, and in other earlier posts.

"Fair" and "unfair" are tricky words to nail down.

I think there are a wide range of factors that explain why HLI has been treated differently than other orgs --  some "fair" under most definitions of the word, some less so. Some of those reasons are adjacent to questions of funding and influence, but I'm not sure they provide much room to criticize HLI's critics.

  1. HLI is running in a lane -- global health/development/wellbeing -- where the evidentiary standards are much higher than in longtermist areas. Part of this is the nature of the work; asking a biosecurity program how many pandemics it has prevented is not workable. Part of it is that there is a very well-funded organization that has been doing CEAs that conensus views as high-quality. Yet another aspect is that GHDW work has been much more limited by funding constraints, which has incentivized GHDW funders to adopt higher standards.
    1. I think people generally need to be kinder to smaller-scale, early-stage efforts . . . but see point 3 below.
  2. HLI is a charity recommender, a significant portion of whose focus currently involves making recommendations to ordinary people (not megadonors, foundations, etc.) I do think the level of scrutiny should ordinarily be higher for charity recommenders, especially those making recommendations to the general public. The purpose of a charity recommender is to evaluate the relative merits of various charities, and for ordinary donors their recommendations may be seen as near-authoritative. A sense that the community needs to carefully scrutinize the recommender's work destroys much of a recommender's value proposition in the first place. And while it's not very utilitarian of me, I do feel more protective of small donors who don't have an in-house staff to pick up on a recommender's mistakes.
  3. I think an overconfident marketing campaign in 2022 did play a major role in how much grace people are willing to extend on the CEA. I haven't been around that long, but this does seem to significantly distinguish HLI from other orgs. I believe that HLI has expressed regret for certain statements, but a framework that compares statements made at that time (that have not been clearly and explicitly retracted) to what the data actually support strikes me as on the "fair" side of the ledger.
  4. This was HLI's first major recommendation; people would be less prone to draw negative inferences about (e.g.) an org whose first four analyses/recommendations were fine but whose fifth had some significant issues. 
  5. StrongMinds spends (and could potentially fundraise) enough money to make a significant dive into its cost-effectiveness worthwhile for critics, but probably not so much as to justify an airtight multi-million dollar workup (including by commissioning our own studies to fill any major holes in the data that would have a big effect on the CEA). So it's an awkward-size program to evaluate.
  6. Pretty much all skeptical analysis is done by volunteers on their own time, and so the volume/quality of that work will heavily depend on who is interested in and available to doing it. It's plausible to me that having a controversial and/or novel framework could motivate more critics to volunteer for duty. 
  7. There could also be a snowball effect; the detection of one significant weakness in a CEA may motivate others to start looking.
  8. HLI asked Forum users to contribute money. Although I take a wide stance on "standing" to criticize organizations, one could reasonably characterize asking users for action as opening the door to some extent. Having an active fundraising ask may also provide a more concrete payoff/impact for criticism, by preventing users from taking an action the critic found undesirable.
  9. HLI has been unusually transparent with data and responsive to criticism, which has made such criticism easier and kept it up longer. I think you're right to be concerned about the ferocity of criticism disincentivizing trransparency and opennness on the margin. 
  10. The barriers to criticizing HLI are much lower. Because HLI has little power, no one is concerned about blowback. Compare that to the recent Omega criticisms of AI labs, which were posted psuedonymously and which had to rely on undisclosed data. Criticism from established community members who sign their work and can show their work carries more weight, and there's a disincentive to writing anonymous criticism (you'll never get any credit for it).

Several of these points are at least adjacent to questions of funding and power, and they cumulatively make me feel at least somewhat uncomfortable, e.g.:

  • It's unlikely an organization with more secure funding would have made a fundraising appeal at this time. Rather, it likely would have laid low until it had produced a new CEA for SM and until more time had passed since the prior harsh posts.
  • HLI may have felt pressure to be particularly transparent and responsive than a more established org. It's unlikely HLI would have been taken seriously if it didn't show its receipts, and it doesn't have the power/prestige needed for a "no real comment" approach to criticism to have a good shot at working. 

That being said, I find it challenging to assign much fault for those factors to the Forum user community on those. For example, in point 10, the unfairness is not that HLI is being criticized by named users who have built up a reputation, but that the criticism of other orgs is disincentivized and psuedonymous. 

I think you're right that the response to HLI may discourage transparency and responsiveness on the margin, and that this is a problem. As a practical matter, I think there are two factors that mitigate this to some extent. One is that I think the criticism of HLI reflects a convergence of a number of factors as listed above, and I'm not sure how much marginal effect comes from their good transparency and responsiveness. Second, I think any startup org trying to pursue HLI-like goals has to be transparent and responsive to get a hearing from the community, so I think it less likely that knowledge of current events will change another org's stance to a materially less open and responsive one.

I'm undecided on the net effect of all of this. My hope is that it will ultimately result in adoption of better epistemic safeguards and communications management -- both at HLI and elsewhere in the ecosystem. (Cf. my recent post on the HLI thread). That would be a good result, although I'd still wish we had gotten there with a lot less rancor.

Wanted to give a shoutout to Ajeya Cotra (from OpenPhil), for her great work explaining AI stuff on a recent Freakonomics podcast series. Her explanations about both her work on the development of AI, and her easy to understand predictions of how AI might progress from here were great, she was my favourite expert on the series.

People have been looking for more high quality public communicators to get EA/AI safety stuff out there, perhaps Ajeya could be a candidate if she's keen?
 

Ajeya is already doing that with Kelsey Piper over at their blog Planned Obsolescence :) 

The net net effect could be positive or negative. Let me untangle it for you.

In favour of net positivity is the net positive human lives saved through net negative, negative net effects on mosquitos and malaria causing a net positive effect on humans. The insecticides positively embed these bed net, net positive effects.

However there’s a catch in favour of net negativity - as nets are positively dragged to net fish, the netted fish suffer net negative net negativity.

While caught between the net positive net positivity to humans and net negative net negativity to fish, we are enmeshed in a net of uncertainty as we struggle to net a clear positive net answer.

Hopefully I didn’t tie you in knots as I strung you along.

In response to @David Mathers ;).

Although I'm enamoured as a mostly-neartermist that the front page (for the first time in my experience) is devoid of AI content, I really would like to hear the job experience and journey of a few AI safety/policy workers for this jobs week. The first 10ish wonderful people who shared are almost all neartermist focused, which probably doesn't represent the full experience of the community.

I'm genuinely interested to understand how your AI safety job works and how you wonderful people motivate yourselves on a day to day basis, when seeing clear progress and wins must be hard a lot of the time. I find it hard enough some days working in Global health!

Or maybe your work is so important, neglected and urgent that your can't spare a couple of hours to write a post ;).



 

Pro tip: you can adjust your settings so you never have to see AI content on the front page

I've been on the forum for maybe 9 months now, and I've been intrigued by the idea of "hits based" giving, explained well in this 2016 article by Holden Karnofsky. The idea that "we will sometimes bet on ideas that contradict conventional wisdom, contradict some expert opinion, and have little in the way of clear evidential support." 

1) Is there a database with a list of donations considered "hits based"by Openphil? If not that would be a helpful and transparent way of tracking success on these. I had a quick look through their donations but its not clear which ones are considered "hits based"

2) Are there any donations considered "hits based" since that 2016 post which have clearly turned out really well, i.e. been winners? We might have seen some these hit based donations resolve positively or negatively, although 6 years is a short timeline so most would remain unresolved. Is there any official or unofficial info on that?

Thanks!


 

Curated and popular this week
Relevant opportunities