New & upvoted

Customize feedCustomize feed
NEW
CommunityCommunity
Personal+

Posts tagged community

Quick takes

Show community
View more
I worked at OpenAI for three years, from 2021-2024 on the Alignment team, which eventually became the Superalignment team. I worked on scalable oversight, part of the team developing critiques as a technique for using language models to spot mistakes in other language models. I then worked to refine an idea from Nick Cammarata into a method for using language model to generate explanations for features in language models. I was then promoted to managing a team of 4 people which worked on trying to understand language model features in context, leading to the release of an open source "transformer debugger" tool. I resigned from OpenAI on February 15, 2024.
56
tlevin
4d
3
I think some of the AI safety policy community has over-indexed on the visual model of the "Overton Window" and under-indexed on alternatives like the "ratchet effect," "poisoning the well," "clown attacks," and other models where proposing radical changes can make you, your allies, and your ideas look unreasonable. I'm not familiar with a lot of systematic empirical evidence on either side, but it seems to me like the more effective actors in the DC establishment overall are much more in the habit of looking for small wins that are both good in themselves and shrink the size of the ask for their ideal policy than of pushing for their ideal vision and then making concessions. Possibly an ideal ecosystem has both strategies, but it seems possible that at least some versions of "Overton Window-moving" strategies executed in practice have larger negative effects via associating their "side" with unreasonable-sounding ideas in the minds of very bandwidth-constrained policymakers, who strongly lean on signals of credibility and consensus when quickly evaluating policy options, than the positive effects of increasing the odds of ideal policy and improving the framing for non-ideal but pretty good policies. In theory, the Overton Window model is just a description of what ideas are taken seriously, so it can indeed accommodate backfire effects where you argue for an idea "outside the window" and this actually makes the window narrower. But I think the visual imagery of "windows" actually struggles to accommodate this -- when was the last time you tried to open a window and accidentally closed it instead? -- and as a result, people who rely on this model are more likely to underrate these kinds of consequences. Would be interested in empirical evidence on this question (ideally actual studies from psych, political science, sociology, econ, etc literatures, rather than specific case studies due to reference class tennis type issues).
2
JWS
1h
0
Going to quickly share that I'm going to take a step back from commenting on the Forum for the foreseeable future. There are a lot of ideas in my head that I want to work into top-level posts to hopefully spur insightful and useful conversation amongst the community, and while I'll still be reading and engaging I do have a limited amount of time I want to spend on the Forum and I think it'd be better for me to move that focus to posts rather than comments for a bit.[1] If you do want to get in touch about anything, please reach out and I'll try my very best to respond. Also, if you're going to be in London for EA Global, then I'll be around and very happy to catch up :) 1. ^ Though if it's a highly engaged/important discussion and there's an important viewpoint that I think is missing I may weigh in
Not sure how to post these two thoughts so I might as well combine them. In an ideal world, SBF should have been sentenced to thousands of years in prison. This is partially due to the enormous harm done to both FTX depositors and EA, but mainly for basic deterrence reasons; a risk-neutral person will not mind 25 years in prison if the ex ante upside was becoming a trillionaire. However, I also think many lessons from SBF's personal statements e.g. his interview on 80k are still as valid as ever. Just off the top of my head: * Startup-to-give as a high EV career path. Entrepreneurship is why we have OP and SFF! Perhaps also the importance of keeping as much equity as possible, although in the process one should not lie to investors or employees more than is standard. * Ambition and working really hard as success multipliers in entrepreneurship. * A career decision algorithm that includes doing a BOTEC and rejecting options that are 10x worse than others. * It is probably okay to work in an industry that is slightly bad for the world if you do lots of good by donating. [1] (But fraud is still bad, of course.) Just because SBF stole billions of dollars does not mean he has fewer virtuous personality traits than the average person. He hits at least as many multipliers than the average reader of this forum. But importantly, maximization is perilous; some particular qualities like integrity and good decision-making are absolutely essential, and if you lack them your impact could be multiplied by minus 20.     [1] The unregulated nature of crypto may have allowed the FTX fraud, but things like the zero-sum zero-NPV nature of many cryptoassets, or its negative climate impacts, seem unrelated. Many industries are about this bad for the world, like HFT or some kinds of social media. I do not think people who criticized FTX on these grounds score many points. However, perhaps it was (weak) evidence towards FTX being willing to do harm in general for a perceived greater good, which is maybe plausible especially if Ben Delo also did market manipulation or otherwise acted immorally. Also note that in the interview, SBF didn't claim his donations offset a negative direct impact; he said the impact was likely positive, which seems dubious.
Trump recently said in an interview (https://time.com/6972973/biden-trump-bird-flu-covid/) that he would seek to disband the White House office for pandemic preparedness. Given that he usually doesn't give specifics on his policy positions, this seems like something he is particularly interested in. I know politics is discouraged on the EA forum, but I thought I would post this to say: EA should really be preparing for a Trump presidency. He's up in the polls and IMO has a >50% chance of winning the election. Right now politicians seem relatively receptive to EA ideas, this may change under a Trump administration.

Popular comments

Recent discussion

JWS posted a Quick Take 1h ago

Going to quickly share that I'm going to take a step back from commenting on the Forum for the foreseeable future. There are a lot of ideas in my head that I want to work into top-level posts to hopefully spur insightful and useful conversation amongst the community, and while I'll still be reading and engaging I do have a limited amount of time I want to spend on the Forum and I think it'd be better for me to move that focus to posts rather than comments for a bit.[1]

If you do want to get in touch about anything, please reach out and I'll try my very best to respond. Also, if you're going to be in London for EA Global, then I'll be around and very happy to catch up :)

  1. ^

    Though if it's a highly engaged/important discussion and there's an important viewpoint that I think is missing I may weigh in

Continue reading
JWS commented on My Lament to EA 1h ago
101
6

Edit: so grateful and positively overwhelmed with all the responses!

I am dealing with repetitive strain injury and don’t foresee being able to really respond to many comments extensively (I’m surprised with myself that I wrote all of this without twitching forearms lol!...

Continue reading

Like others, I just want to say I'm so sorry that you had this experience. It isn't one I recognise from my own journey with EA, but this doesn't invalidate what you went through and I'm glad you're moving in a direction that works for you as a person and your values. You are valuable, your life and perspective is valuable, and I wish all you all the best in your future journey.

Indirectly, I'm going to second @Mjreard below - I think EA should be seen as beyond a core set of people and institutions. If you are still deeply driven by the ideals EA was inspi... (read more)

5
Denkenberger
4h
I'm sorry to hear about your negative experiences in EA. Sorry also to hear that - have you tried voice recognition software? It was a game changer for me back in 2000 (and it's gotten a lot better since then!) - both for RSI and productivity.  
10
jessica_mccurdy
12h
Hey I am really sorry to hear about all of these negative experiences.  I feel lucky to have gotten to work with you over the years and seen the positive impact you have had on others in the community and the exciting work you have moved into. I think the community will be losing a really lovely person. I admire both your courage in posting this and that you are prioritizing your well-being right now.  I was sad to hear that our team contributed to your negative experiences though I definitely understand. When I first was introduced to the idea of focusing on top universities, I also felt quite uncomfortable with the implications. I knew so many brilliant people not from those top universities and knew many systemic disadvantages prevent people from attending them.  That being said, I do still endorse this prioritization (though not an exclusive one!). In case it’s useful context, for you or others reading this post, I’ve written some reasons why I think this. When piloting new forms of support or being limited in the number of universities a small team can give higher touch support to, hard decisions have to be made about which universities get that support. I believe most of the people who go on to have incredibly impactful careers will come from outside those top universities. That is part of why we set up UGAP - to help provide support and opportunities to promising individuals and groups around the world who might not have the same access and advantages. However, top universities are the places with the highest concentrations of people who ultimately have a very large influence on the world. Partially that is because of the screening mechanisms of the university and but partially it is because of the built-in benefits and unfair advantages that students receive while they attend those universities. Given that this is the reality we're working with, I believe we should do our best to leverage the existing system and opportunities at these universities to h
1
0

TL;DR: We need technology and infrastructure specialists. You don't need to specialise in cybersecurity to have an impactful career addressing AI risk, or even to improve cybersecurity.

I've been providing advice and mentoring to EAs on cybersecurity and IT careers for a few years now, mostly at conferences. I've regularly made the case (often to the relief of the mentee) that people on cybersecurity and other IT career pathways should consider staying the course rather than retraining as machine learning researchers.

This year, after increased community focus on information security in relation to AI risk, I am now often asked how to retrain into a cybersecurity specialisation. In response, I'm making the case that an oversupply of cybersecurity professionals is not optimal, and (possibly un-intuitively) is not ideal for cybersecurity either.

Organisations working on AI risk, as well as other...

Continue reading
Sign up for the Forum's email digest
You'll get a weekly email with the best posts from the past week. The Forum team selects the posts to feature based on personal preference and Forum popularity, and also adds some announcements and a classic post.

Summary

  1. Where there’s overfishing, reducing fishing pressure or harvest rates — roughly the share of the population or biomass caught in a fishery per fishing period — actually allows more animals to be caught in the long run.
  2. Sustainable fishery management policies
...
Continue reading

Thanks for tagging me! I'll read the post and your comment with care.

This is a linkpost for https://ailabwatch.org

I'm launching AI Lab Watch. I collected actions for frontier AI labs to improve AI safety, then evaluated some frontier labs accordingly.

It's a collection of information on what labs should do and what labs are doing. It also has some adjacent resources, including a list...

Continue reading
3
Dan H
5h
OpenAI has made a hard commitment to safety by allocating 20% compute (~20% of budget) for the superalignment team. That is a huge commitment which isn't reflected in this.

I agree such commitments are worth noticing and I hope OpenAI and other labs make such commitments in the future. But this commitment is not huge: it's just "20% of the compute we've secured to date" (in July 2023), to be used "over the next four years." It's unclear how much compute this is, and with compute use increasing exponentially it may be quite little in 2027. Possibly you have private information but based on public information the minimum consistent with the commitment is quite little.

It would be great if OpenAI or others committed 20% of their compute to safety! Even 5% would be nice.

Does anyone know the latest estimate of what percentage of US/Western charity goes to poverty broadly and international poverty specifically?

A 2013 Dylan Matthews piece in WaPo cites a 2007 estimate. Googling isn't helping much. 

I may be writing an opinion piece for...

Continue reading

Perhaps it’s in the Giving USA Annual Survey? You need a subscription to access it though.

I am writing this post in response to a question that was raised by Nick a few days ago,

1) as to whether the white sorghum and cassava that our project aims to process will be used in making alcohol, 2) whether the increase in production of white sorghum and cassava...

Continue reading

Hi Roddy, thanks very much too for this message. Here are my answers to these questions:

1). Currently, our farmers are within a radius of about 40km from the UCF, and because the volume of their sorghum is still a bit small, it's the UCF team itself that gathers all these farmers' sorghum using a motorbike, and brings it to the UCF, from where we take it to Kampala. 

We also have a dump truck at the UCF, and whenever the load we are going to carry is a bit big, we use this truck instead. Right now, all these costs (fuel, transport etc) are covered by t... (read more)

DC commented on Why I'm doing PauseAI 6h ago

GPT-5 training is probably starting around now. It seems very unlikely that GPT-5 will cause the end of the world. But it’s hard to be sure. I would guess that GPT-5 is more likely to kill me than an asteroid, a supervolcano, a plane crash or a brain tumor. We can predict...

Continue reading
DC
6h2
0
0

I know of one that is less widely reported; not sure if they're counted in the two Joseph Miller knows of that are less widely reported, or if separate.

This announcement was written by Toby Tremlett, but don’t worry, I won’t answer the questions for Lewis.

Lewis Bollard, Program Director of Farm Animal Welfare at Open Philanthropy, will be holding an AMA on Wednesday 8th of May. Put all your questions for him on this thread...

Continue reading

What percent of farmed animal welfare advocacy does Open Phil fund? How valuable would it be for the cause area to have additional major funders?

4
James Özden
10h
In your view, what are some of the biggest challenges facing the farmed animal movement today and what is Open Phil doing about them?
3
James Özden
12h
What are some important lessons or things you've learned on how to do grantmaking well over the past 9 years that you would give to yourself when you were starting at OP?