MathiasKB

Director @ Center for Effective Aid Policy
4975 karmaJoined aidpolicy.org

Comments
237

I think there's a lot of local maxima that are very juicy. I would encourage people to look at the opportunities around them that others would miss, and try my best to foster a culture that helps its members discover them.

A great example of someone doing this is Abdurrahman who took the initiative to create EA in Arabic which I expect will be really impactful. I don't expect there were many EA jobs available to him in Saudi, but he looked around and found a gap (no resources on EA in one of the world's biggest languages) and executed on the opportunity.

I am currently looking into an animal welfare intervention which South American EA's would be much better suited to do than anyone else. Some time ago I looked into policy interventions to improve the water sanitation efforts of the Jal Jeevan Mission in India. An Indian EA from the right state would be much better suited to carry out the sanitation advocacy for JJM than I am.

I've yet to find a region of the world without opportunities, but most of them won't be listed in a career guide!

I'm grappling with this exact issue. I think AI is the most important technology humanity will event, but I'm skeptical of the EV of much work on the technology. Still it seems that it should be the only reasonable thing to spend all my time thinking about, but even then I'm not sure I'd arrive at anything useful.

And the opportunity cost is saving hundreds of lives. I don't think there is any other question that has cost me as much sleep as this one.

forecasting newsletter by nuno sempere

Excerpt from the most recent update from the ALERT team:

 

Highly pathogenic avian influenza (HPAI) H5N1: What a week! The news, data, and analyses are coming in fast and furious.

Overall, ALERT team members feel that the risk of an H5N1 pandemic emerging over the coming decade is increasing. Team members estimate that the chance that the WHO will declare a Public Health Emergency of International Concern (PHEIC) within 1 year from now because of an H5N1 virus, in whole or in part, is 0.9% (range 0.5%-1.3%). The team sees the chance going up substantially over the next decade, with the 5-year chance at 13% (range 10%-15%) and the 10-year chance increasing to 25% (range 20%-30%).

 

their estimated 10 year risk is a lot higher than I would have anticipated.

I suspect the primary reasons you want to break up Deepmind from Google is to:

  1. Increase their autonomy, reducing pressure from google to race
  2. Reduce Deepmind's access to capital and compute, reducing their competitiveness

Perhaps that goes without saying, but I think it's worth explicitly mentioning. In a world without AI risk, I don't believe you would be citing various consumer harms to argue for a break up.

The traditional argument for breaking up companies and preventing mergers is to reduce the company's market power, increasing consumer surplus. In this case, the implicit reason for breaking up Deepmind is to decrease its competitiveness thus reducing consumer surplus.

I think it's perfectly fine to argue for this, I just really want us to be explicit about it.

I'm awestruck, that is an incredible track record. Thanks for taking the time to write this out.

These are concepts and ideas I regularly use throughout my week and which have significantly shaped my thinking. A deep thanks to everyone who has contributed to FHI, your work certainly had an influence on me.

I think I'm sympathetic to Oxford's decision.

By the end, the line between genuine scientific inquiry and activistic 'research' got quite blurry at FHI. I don't think papers such as: 'Proposal for a New UK National Institute for Biological Security', belong in an academic institution, even if I agree with the conclusion.

One thing that stood out to me reading the comments on Reddit, was how much of the poor reception that could have been avoided with a little clearer communication.

For people such as MacAskill, who are deeply familiar with effective altruism, the question: "Why would SBF pretend to be an Effective Altruist if he was just looking to do fraud?"  is quite the conundrum. Of all the types of altruism, why specifically pick EA as the vehicle to smuggle your reputation? EA was already unlikeable and elitist before the scandal. Why not donate to puppies and Harvard like everyone else?

I actually admire MacAskill for asking that question. The easy out, would be to say: "how could we have been so foolish, SBF was clearly never a real EA". But he instead grapples with the fact that SBF seems to have been genuinely motivated by effective altruism, and that these ideals must have played some part in SBFs decision to commit fraud.

But for any listener who is not as deeply familiar with the effective altruism movement, and doesn't know its reputation, the question comes off as hopelessly naive. The emphasis they hear is: "Why would SBF, a fraudulent billionaire, pretend to be an Effective Altruist?" The answer to that is obvious - malicious actors pretend to be altruistic all the time!

I see EA communication make this mistake all the time. A question or idea whose merit is obvious to you might not be obvious to everyone else if you don't spell out the assumptions it rests on.

I think I am misunderstanding the original question then?

I mean if you ask: "what you all think about the series as an entry point for talking about some of these EA issues with friends, family, colleagues, and students"

then the reach is not the 10 million people watching the show, it's the people you get a chance to speak to.

Load more