I met Australia’s Assistant Minister for Defence last Friday. I asked him to write an email to the Minister in charge of AI, asking him to establish an AI Safety Institute. He said he would. He also seemed on board with not having fully autonomous AI weaponry.
All because I sent one email asking for a meeting + had said meeting.
Advocacy might be the lowest hanging fruit in AI Safety.
In May and June of 2023, I (Akash) had about 50-70 meetings about AI risks with congressional staffers. …
In March of 2023, I started working on some AI governance projects at the Center for AI Safety. One of my projects involved helping CAIS respond to a Request for Comments about AI Accountability that was released by the NTIA.
As part of that work, I started thinking a lot about what a good regulatory framework for frontier AI would look like. For instance: if I could set up a licensing regime for frontier AI systems, what would it look like? Where in the US government would it be housed? What information would I want it to assess?
I began to wonder how actual policymakers would react to these ideas. I was also curious to know more about how policymakers were thinking about AI extinction risks and catastrophic risks.
I started asking other folks in AI Governance. The vast majority had not talked to congressional staffers (at all). A few had experience talking to staffers but had not talked to them about AI risk. A lot of people told me that they thought engagement with policymakers was really important but very neglected. And of course, there are downside risks, so you don’t want someone doing it poorly.
After consulting something like 10-20 AI governance folks, I asked CAIS if I could go to DC and start talking to congressional offices. The goals were to (a) raise awareness about AI risks, (b) get a better sense of how congressional offices were thinking about AI risks, (c) get a better sense of what kinds of AI-related priorities people at congressional offices had, and (d) get feedback on my NTIA request for comment ideas.
CAIS approved, and I went to DC in May-June 2023. And just to be clear, this wasn’t something CAIS told me to do– this was more of an “Akash thing” that CAIS was aware was happening.
Like you, Akash just cold-emailed people:
I sent a mass email to tech policy staffers, and I was pretty impressed by the number who responded. The email was fairly short, mentioned that I was at CAIS, had 1-2 bullets about what CAIS does, and had a bullet point about the fact that I was working on an NTIA request for comment.
I think it was/is genuinely the case that Congressional staffers are extremely interested in AI content right now. Like, I don’t think I would’ve been able to have this many meetings if I was emailing people about other issues.
There’s a lot of concrete learnings in that writeup; definitely worth reading I think.
I don’t know how to say this in a way that won’t come off harsh, but I’ve had meetings with > 100 people in the last 6 months in AI Safety and I’ve been surprised how poor the industry standard is for people:
- Planning meetings (writing emails that communicate what the meeting will be about, it’s purpose, my role) - Turning up to scheduled meetings (like literally, showing up at all) - Turning up on time - Turning up prepared (i.e. with informed opinions on what we’re going to discuss) - Turning up with clear agendas (if they’ve organised the meeting) - Running the meetings - Following up after the meeting (clear communication about next steps)
I do not have an explanation for what is going on here but it concerns me about the movement interfacing with government and industry. My background is in industry and these just seem to be things people take for granted?
It is possible I give off a chill vibe, leading to people thinking these things aren’t necessary, and they’d switch gears for more important meetings!
My 2 cents is that I think this is a really important observation (I think it’s helpful) and I would be interested to see what other peoples feedback is on this front.
hypothesis that springs to mind, might or might not be useful for engaging with it productively. might be wrong depending on what class of people you have been having meetings with.
when you select for people working on AI risks, you select for people who are generally less respective of status quo and social norms. you select for the kind of person who will generally do something less, because of the reason that its a social norm. in order for this reference class of person to do a thing, they kind of who through their personal reasoning methods reaches the conclusion “It would be worth the effort for me to change myself to become the kind of person who shows up on time consistently, compared to other things I could be spending my effort on”. they might just figure its a better use of their effort to think about their research all day.
I would guess that most people on this earth don’t show up on time because they reasoned it through that this is a good idea, they do it because it has been drilled into them through social norms, and they valued those social norms higher.
(note: this comment is not intended to be an argument that showing up on time is a waste of time)
When AI Safety people are also vegetarians, vegans or reducetarian, I am pleasantly surprised, as this is one (of many possible) signals to me they’re “in it” to prevent harm, rather than because it is interesting.
I am skeptical that exhibiting one of the least cost-effective behaviors for the purpose of reducing suffering would be correlated much with being “in it to prevent harm”. Donating to animal charities, or being open to trade with other people about reducing animal suffering, or other things that are actually effective seem like much better indicator that people are actually in it to prevent harm.
My best guess is that most people who are vegetarians, vegans or reducetarians, and are actually interested in scope-sensitivity, are explicitly doing so for signaling and social coordination reasons (which to be clear, I think has a bunch going for it, but seems like the kind of contingent fact that makes it not that useful as a signal of “being in it to prevent harm”).
Have you actually just talked to people about their motives for reducing their animal product consumption, or read about them? Do you expect them to tell you it’s mostly for signaling or social coordination, or (knowingly or unknowingly) lie that it isn’t?
I’d guess only a minority of EAs go veg mostly for signaling or social coordinaton.
Other reasons I find more plausibly common, even among those who make bigger decisons in scope-sensitive ways:
Finding it cost-effective in an absolute sense, e.g. >20 chickens spared per year, at small personal cost (usually ignoring opportunity costs to help others more). (And then maybe falsely thinking you should go veg to do this, rather than just cut out certain animal products, or finding it easier to just go veg than navigate exceptions.)
Wanting to minimize the harm they personally cause to others, or to avoid participating in overall harmful practices, or deontological or virtue ethics reasons, separately from actively helping others or utilitarian-ish reasons.
Not finding not eating animal products very costly, or not being very sensitive to the costs. It may be easy or take from motivational “budgets” psychologically separate from work or donations.
Finding eating animal products psychologically or time costly, e.g. just feeling guilty and worrying about it, cognitive dissonance.
Seeing so many other EAs veg, and this making it seem very impactful to be veg for animals (whether or not it actually is).
As a way to live out their commitment daily that other animals matter and increase/maintain their concern for them (and/or other vulnerable moral patients easy to ignore).
To get other people around them interested in helping animals.
I’d guess most EAs who go veg or reducetarian do so for reasons like 1, 2, 3, 4 and/or 5. 1, 2, 3 and 4 are why I went vegan, around when I first got into EA. I also used to make arguments like 1, just comparing the harms to animals and your personal costs (after the drowning child thought experiment) when I was an EA organizer a few years ago.
I don’t think the opportunity costs of veg diets are on people’s minds much, so it seems likely that people are ignoring them, either for bad reasons or because they don’t seem important enough to them.
I’m still (bival)vegan, but mostly for reasons like 3, 6, just preferring not to eat animal products anymore, and maybe to meet (my perception of) others’ expectations of me.
I’m not even convinced veganism is better for animals anymore through its supply and demand effects, when you include the effects on wild animals (mostly wild arthropods and fish).
I have had >30 conversations with EA vegetarians and vegans about their reasoning here. The people who thought about it the most seem to usually settle on it for signaling reasons. Maybe this changed over the last few years in EA, but it seemed to be where most people I talked to where at when I had lots of conversations with them in 2018.
I agree that many people say (1), but when you dig into it it seems clear that people incur costs that would be better spent on donations, and so I don’t think it’s good reasoning. As far as I can tell most people who think about it carefully seem to stop thinking its a good reason to be vegan/vegetarian.
I do think the self-signaling and signaling effects are potentially substantial.
I also think (4) is probably the most common reason, and I do think probably captures something important, but it seems like a bad inference that “someone is in it to prevent harm” if (4) is their reason for being vegetarian or vegan.
when you dig into it it seems clear that people incur costs that would be better spent on donations, and so I don’t think it’s good reasoning.
I’ve thought a lot about this, because I’m serious about budgeting and try to spend as little money as possible to make more room for investments and donations. I also have a stressful job and don’t like to spend time cooking. I did not find it hard to switch to being vegan while keeping my food budget the same and maintaining a high-protein diet.
Pea protein is comparable in cost per gram protein to the cheapest animal products where I live, and it requires no cooking. Canned beans are a bit more expensive but still very cheap and also require no cooking. Grains, ramen and cereal are very cheap sources of calories. Plant milks are more expensive than cow’s milk, but still fit in my very low food budget. Necessary supplements like B12 are very cheap.
On a day-to-day basis, the only real cooking I do is things like pasta, which don’t take much time at all. I often go several weeks without doing any real cooking. I’d bet that I both spend substantially less money on food and substantially less time cooking than the vast majority of omnivores, while eating more protein.
As a vegan it’s also easy to avoid spending money in frivolous ways, like on expensive ready-to-eat snacks and e.g. DoorDash orders.
I haven’t had any health effects either, or any differences in how I feel day-to-day, after more than 1 year. Being vegan may have helped me maintain a lower weight after I dropped several pounds last year, but it’s hard to know the counterfactual.
I didn’t know coming in that being vegan would be easy; I decided to try it out for 1 month and then stuck with it when I “learned” how to do it. There’s definitely a learning curve, but I’d say that for some people who get the hang of it, reason (1) in Michael’s comment genuinely applies.
I don’t think one year is enough time to observe effects. Anecdotically, I think (but am not sure) that I started to have problems after three years of being a vegetarian.
I agree that many people say (1), but when you dig into it it seems clear that people incur costs that would be better spent on donations, and so I don’t think it’s good reasoning.
Do you mean financial costs, or all net costs together, including potentially through time, motivation, energy, cognition? I think it’s reasonably likely that for many people, there are ~no real net (opportunity) costs, or that it’s actually net good (but if net good in those ways, then that would probably be a better reason than 1). Putting my thoughts in a footnote, because they’re long and might miss what you have in mind.[1]
I also think (4) is probably the most common reason, and I do think probably captures something important, but it seems like a bad inference that “someone is in it to prevent harm” if (4) is their reason for being vegetarian or vegan.
Ya, that seems fair. If they had the option to just stop thinking and feeling bad about it and chose that over going veg, which is what my framing suggests they would do, then it seems the motivation is to feel better and get more time, not avoid harming animals through their diets. This would be like seeing someone in trouble, like a homeless person, and avoiding them to avoid thinking and feeling bad about them. This can be either selfish or instrumentally other-regarding, given opportunity costs.
If they thought (or felt!) the right response to the feelings is to just go veg and not to just stop thinking and feeling bad about it, then I would say they are in it to prevent harm, just guided by their feelings. And their feelings might not be very scope-sensitive, even if they make donation and career decisions in scope-sensitive ways. I think this is kind of what virtue ethics is about. Also potentially related: “Do the math, then burn the math and go with your gut”.
Financial/donations: It’s not clear to me that my diet is more expensive than if I were omnivorous. Some things I’ve substituted for animal products are cheaper and others are more expensive. I haven’t carefully worked through this, though. It’s also not clear that if it were more expensive, that I would donate more, because of how I decide how much to donate, which is based on my income and a vague sense of my costs of living, which probably won’t pick up differences due to diet (but maybe it does in expectation, and maybe it means donating less later, because I’ll have less money to donate later). If I budgeted more precisely, that would have time costs (which might not come out of work, anyway, though). And if I weren’t vegan, maybe I’d be less motivated to donate as much (although this is more like “self-signaling”, or altruism sharpens altruism).
Time: I doubt most veg*ns would work more hours (allowing more impactful work or more donations) if they weren’t managing or accommodating their veg*n diet. Time spent on the diet is small and probably doesn’t really come out of work hours. But this can depend on their particular circumstances. Maybe someone gets their blood tested more frequently because of their diet, and this specifically comes out of their work hours.
Time costs might also be incurred by others, like event or working space organizers (like you, as I think you’ve brought up before), and veg*ns may not appreciate this, or might (I’d guess correctly) think one more vegan makes ~no difference. Maybe it would be better if everyone agreed to eat whatever at events, and the animal products were offset by animal charity donations, or that time was just given back to organizers to work on other important things without any specific donation offset.
Effects on energy and cognition will vary by person. I think there are risks here people should make some effort to minimize (e.g. blood testing, supplements). That effort can come out of their time and donations, but that’s already accounted for above. There might be some remaining (expected) costs even after this. Or there could be net benefits, in case they end up with an overall healthier diet this way (and might not be motivated to do so except to go veg; it’s easier to avoid unhealthy foods you also object to morally).
Do you mean financial costs, or all net costs together, including potentially through time, motivation, energy, cognition?
I meant net costs all together, tough I agree that if you take into account motivation “net costs” becomes a kind of tricky concept, and many people can find it motivating, and that is important to think about, but also really doesn’t fit nicely into a harm-reducing framework.
Financial/donations: It’s not clear to me that my diet is more expensive than if I were omnivorous. Some things I’ve substituted for animal products are cheaper and others are more expensive.
I mean, being an onmivore would allow you to choose between more options. Generally having more options very rarely hurts you.
many people can find it motivating, and that is important to think about, but also really doesn’t fit nicely into a harm-reducing framework.
Ya, I guess the value towards harm reduction would be more indirect/instrumental in this case.
I mean, being an onmivore would allow you to choose between more options. Generally having more options very rarely hurts you.
I think this is true of idealized rational agents with fixed preferences, but I’m much less sure about actual people, who are motivated in ways they wouldn’t endorse upon reflection and who aren’t acting optimally impartially even if they think it would be better on reflection if they did.
By going veg, you eliminate or more easily resist the motivation to eat more expensive animal products that could have net impartial opportunity costs. Maybe skipping (expensive) meats hurts you in the moment (because you want meat), but it saves you money to donate to things you think matter more. You’d be less likely to correctly — by reflection on your impartial preferences — skip the meat and save the money if you weren’t veg.
And some people are not even really open to (cheaper) plant-based options like beans and tofu, and that goes away going veg. That was the case for me. My attitude before going veg would have been irrational from an impartial perspective, just considering the $ costs that could be donated instead.
Of course, some people will endorse being inherently partial to themselves upon reflection, so eating animal products might seem fine to them even at greater cost. But the people inclined to cut out animal products by comparing their personal costs to the harms to animals probably wouldn’t end up endorsing their selfish motivation to eat animal products over the harms to animals.
The other side is that a veg*n is more motivated to eat the more expensive plant-based substitutes and go to vegan restaurants, which (in my experience) tend to be more expensive.
I’m not inclined to judge how things will shake out based on idealized models of agents. I really don’t know either way, and it will depend on the person. Cheap veg diets seem cheaper than cheap omni diets, but if people are eating enough plant-based meats, their food costs would probably increase.
Here are prices in Canadian $/kg of protein, for the versions of foods that seemed cheapest per kg of protein from Walmart Canada and Amazon Canada.
And then extra supplements for veg EAs.
For restaurants, as long as you avoid expensive plant-based substitutes and vegan-specific restaurants, I think veg options are usually cheaper. Of course, a veg EA will be tempted by these sometimes, too!
To be clear, I don’t think it’s that important to minimize the cost of your diet. Things like rent, vehicle, travel, and how often you eat out (when it doesn’t help you do more work) are probably more important if you want to save money.
I have had >30 conversations with EA vegetarians and vegans about their reasoning here. The people who thought about it the most seem to usually settle on it for signaling reasons. Maybe this changed over the last few years in EA, but it seemed to be where most people I talked to where at when I had lots of conversations with them in 2018.
I guess this is a bit pedantic, but you originally wrote “My best guess is that most people who are vegetarians, vegans or reducetarians, and are actually interested in scope-sensitivity, are explicitly doing so for signaling and social coordination reasons”. I think veg EAs are generally “actually interested in scope-sensitivity”, whether or not they’re thinking about their diets correctly and in scope-sensitive ~utilitarian terms. “The people who thought about it the most” might not be representative, and more representative motivations might be better described as “in it to prevent harm”, even if the motivations turn out to be not utilitarian, not appropriately scope-sensitive or misguided.
I actually think it is a huge risk because when the shit hits the fan I want to know people are optimising for the right thing. It is like the old saying; “it is only a value if it costs you something”
Anyway, I’d like to think that orgs suss this out in the interview stage.
People in EA are ridiculously helpful. I’ve been able to get a Careers Conference off the ground because people in other organisations are happy to help out in significant ways. In particular, Elliot Teperman and Bridget Loughhead from EA Australia, Greg Sadler from the Good Ancestors Project and Chris Leong. Several people have also volunteered to help at the Conference as well as on other tasks, such as running meetups and creating content.
In theory I’m running an organisation by myself (via an LTFF grant), but because it is nested within the EA/AIS space, I’ve got the support of a big network. Otherwise this whole project would be insane to start.
If someone is serious about doing Movement Building, but worries about doing it alone, this feels like important information.
Another thing you can do is send comments proposed legislation on regulations.gov. I did so last week about a recent californian bill on open-sourcing model weights (now closed). In the checklist (screenshot below) they say: “the comment process is not a vote – one well supported comment is often more influential than a thousand form letters”. There are people much more qualified on AI risk than I over here, so in case you didn’t know, you might want to keep an eye on new regulation coming up. It doesn’t take much time and seems to have a fairly big impact.
I think it is good to have some ratio of upvoted/agreed : downvotes/disagreed posts in your portfolio. I think if all of your posts are upvoted/high agreeance then you’re either playing it too safe or you’ve eaten the culture without chewing first.
I think some kinds of content are uncontroversially good (e.g. posts that are largely informational rather than persuasive), so I think some people don’t have a trade-off here.
[IMAGE] Extremely proud and excited to watch Greg Sadler (CEO of Good Ancestors Policy) and Soroush Pour (Co-Founder of Harmony Intelligence) speak to the Australian Federal Senate Select Committee on Artificial Intelligence. This is a big moment for the local movement. Also, check out who they ran after.
My previous take on writing to Politicians got numbers, so I figured I’d post the email I send below.
I am going to make some updates, but this is the latest version:
---
Hi [Politician]
My name is Yanni Kyriacos, I live in Coogee, just down the road from your electorate.
If you’re up for it, I’d like to meet to discuss the risks posed by AI. In addition to my day job building startups, I do community / movement building in the AI Safety / AI existential risk space. You can learn more about AI Safety ANZ by joining our Facebook group here or the PauseAI movement here. I am also a signatory of Australians for AI Safety—a group that has called for the Australian government to set up an AI Commission (or similar body).
Recently I worked with Australian AI experts (such as Good Ancestors Policy) in making a submission to the recent safe and response AI consultation process. In the letter, we called on the government to acknowledge the potential catastrophic and existential risks from artificial intelligence. More on that can be found here.
There are many immediate risks from already existing AI systems like ChatGPT or Midjourney, such as disinformation or improper implementation in various businesses. In the not-so-distant future, certain safety nets will need to be activated (such as a Universal Basic Income policy) in the event of mass unemployment due to displacement of jobs with robots and AI systems.
But of greatest concern is the speed at which we are marching towards AGI (artificial general intelligence) – systems that will have cognitive abilities at or above human level.
Half of AI researchers believe that there is a 10% or greater chance that the invention of artificial superintelligence will mean the end of humanity. Among AI safety scientists, this chance is estimated to be an average of 30%. And these levels of risk aren’t just a concern for people in the far-distant future, with prediction markets such as Metaculus showing these kinds of AI could be invented in the next term of government.
Notable examples of individuals sounding the alarm are Prof. Geoffrey Hinton and Prof. Yoshua Bengio, both Turing-award winners and pioneers of the deep learning methods that are currently achieving the most success. The existential risk of AI has been acknowledged by hundreds of scientists, the UN, the US and recently the EU.
To make a long story short: we don’t know how to align AI with the complex goals and values that humans have. When a superintelligent system is realised, there is a significant risk it will pursue a misaligned goal without us being able to stop it. And even if such a superhuman AI remains under human control, the person (or government) wielding such a power could use this to drastically, irreversibly change the world. Such an AI could be used to develop new technologies and weapons, manipulate masses of people or topple governments.
The advancements in the AI landscape have progressed much faster than anticipated. In 2020, it was estimated that an AI would pass university entrance exams by 2050. This goal was achieved in March 2023 by the system GPT-4 from OpenAI. These massive, unexpected leaps have prompted many experts to request a pause in AI development through an open letter to major AI companies. The letter has been signed over 33,000 times so far, including many AI researchers and tech figures.
Unfortunately, it seems that companies are not willing to jeopardise their competitive position by voluntarily halting development. A pause would need to be imposed by a government. Luckily, there seems to be broad support for slowing down AI development. A recent poll indicates that 63% of American support regulations to prevent AI companies from building superintelligent AI. At the national level, a pause is also challenging because countries have incentives to not fall behind in AI capabilities. That’s why we need an international solution.
The UK organised an AI Safety Summit on November 1st and 2nd at Bletchley Park. We hoped that during this summit, leaders would work towards sensible solutions that prevent the very worst of the risks that AI poses. As such I was excited to see that Australia signed the The Bletchley Declaration, agreeing that this risk is real and warrants coordinated international action. However, the recent policy statements by Minister Husic don’t seem to align with the urgency that experts are seeing. The last safe moment to act could be very soon.
The Summit has not yet produced an international agreement or policy. We have seen proposals being written by the US Senate, and even AI company CEOs have said there is “overwhelming consensus” that regulation is needed. But no proposal so far has seriously considered ways to slow down or prevent a superintelligent AI from being created. I am afraid that lobbying efforts by AI companies to keep regulation at a minimum are turning out to be highly effective.
It’s essential that the government follows through on its commitment at Bletchley Park to create a national or regional AI safety body. We have such bodies for everything from the risk of plane crashes to the risk of tsunamis. We urgently need one on ensuring the safety of AI systems.
Anyway, I’d love to discuss this more in person or via zoom if you’re in town soon.
Thanks for sharing, Yanni, and it is really cool that you managed to get Australia’s Assistant Minister for Defence interested in creating an AI Safety Institute!
More on that can be found here.
Did you mean to include a link?
In 2020, it was estimated that an AI would pass university entrance exams by 2050.
The Metaculus’ question you link to involves meeting many conditions besides passing university exams:
For these purposes we will thus define “AI system” as a single unified software system that can satisfy the following criteria, all easily completable by a typical college-educated human.
Able to reliably pass a Turing test of the type that would win the Loebner Silver Prize.
Be able to score 75th percentile (as compared to the corresponding year’s human students; this was a score of 600 in 2016) on all the full mathematics section of a circa-2015-2020 standard SAT exam, using just images of the exam pages and having less than ten SAT exams as part of the training data. (Training on other corpuses of math problems is fair game as long as they are arguably distinct from SAT exams.)
Be able to learn the classic Atari game “Montezuma’s revenge” (based on just visual inputs and standard controls) and explore all 24 rooms based on the equivalent of less than 100 hours of real-time play (see closely-related question.)
[PHOTO] I sent 19 emails to politicians, had 4 meetings, and now I get emails like this. There is SO MUCH low hanging fruit in just doing this for 30 minutes a day (I would do it but my LTFF funding does not cover this). Someone should do this!
I would do it but my LTFF funding does not cover this
(Speaking as someone on LTFF, but not on behalf of LTFF)
How large of a constraint is this for you? I don’t have strong opinions on whether this work is better than what you’re funded to do, but usually I think it’s bad if LTFF funding causes people to do things that they think is less (positively) impactful!
We probably can’t fund people to do things that are lobbying or lobbying-adjacent, but I’m keen to figure out or otherwise brainstorm an arrangement that works for you.
RIP to any posts on anything earnest over the last 48 hours. Maybe in future we don’t tag anything April Fools and it is otherwise a complete blackout on serious posts 😅
I think the hilarity is in the confusion / click bait. Your idea would rob us of this! I think the best course of action is for anyone with a serious post to wait until April 3 :|
Not a solution to everything mentioned here- but a reminder that you can click “customize feed” at the top of the page and remove all posts tagged april fools.
I think if you work in AI Safety (or want to) it is very important to be extremely skeptical of your motivations for working in the space. This applies to being skepticism of interventions within AI Safety as well.
For example, EAs (like most people!) are motivated to do things they’re (1) good at (2) see as high status (i.e. people very quietly ask themselves ‘would someone who I perceive as high status approve of my belief or action?‘). Based on this, I am worried that (1) many EAs find protesting AI labs (and advocating for a Pause in general) cringy and/or awkward (2) Ignore the potential impact of organisations such as PauseAI.
We might all literally die soon because of misaligned AI, so what I’m recommending is that anyone seriously considering AI Safety as a career path spends a lot of time on the question of ‘what is really motivating me here?’
fwiw i think this works in both directions—people who are “action” focussed probably have a bias towards advocacy / protesting and underweight the usefulness of research.
I have written 7 emails to 7 Politicians aiming to meet them to discuss AI Safety, and already have 2 meetings.
Normally, I’d put this kind of post on twitter, but I’m not on twitter, so it is here instead.
I just want people to know that if they’re worried about AI Safety, believe more government engagement is a good thing and can hold a decent conversation (i.e. you understand the issue and are a good verbal/written communicator), then this could be an underrated path to high impact.
Another thing that is great about it is you can choose how many emails to send and how many meetings to have. So it can be done on the side of a “day job”.
Big AIS news imo: “The initial members of the International Network of AI Safety Institutes are Australia, Canada, the European Union, France, Japan, Kenya, the Republic of Korea, Singapore, the United Kingdom, and the United States.”
Another shower thought—I’ve been vegetarian for 7 years and vegan for 6.5.
I find it very surprising and interesting how little it has nudged anyone in my friends / family circle to reduce their consumption of meat.
I would go as far as to say it has made literally no difference!
Keep in mind I’m surrounded by extremely compassionate and mostly left leaning people.
I also have relatively wide networks and I normally talk about animal welfare in a non annoying way.
You’d have thought in that time one person would have been nudged—nope!
My intention going vegan wasn’t as some social statement, but if you’d asked me 7 years later would there have been some social effect, I would have guessed yes.
So your surprise/expectation seems reasonable! Of course, I don’t know whether it’s actually surprising, since presumably whether anyone actually converts depends on lots of other features of a given social network (do your networks contain a lot of people who were already vegetarian/vegan?).
Thanks for sharing that data David, I appreciate that :)
Do you think it can answer the question; If you’re the first in your family to switch to veganism/vegetarianism, how likely is it that another family member will follow suit within X years?
(Also, no, there aren’t any vegetarians or vegans among my friends or family).
FWIW I’d probably count myself as having ~ 30 close friends and family.
This also seems like good research for the GWWC pledge.
Do you think it can answer the question; If you’re the first in your family to switch to veganism/vegetarianism, how likely is it that another family member will follow suit within X years?
I’m afraid that. I think we’d need to know more about the respondents’ families and the order in which they adopted veganism/vegetarianism to assess that. I agree that it sounds like an interesting research question!
Have you tried cooking your best vegan recipes for others? In my experience sometimes people ask for the recipe and make it for themselves later, especially health-conscious people. For instance, I really like this vegan pumpkin pie that’s super easy to make: https://itdoesnttastelikechicken.com/easy-vegan-pumpkin-pie/
Thanks for the comment Sean. Because of me they’ve had to (seemingly gladly) eaten vegan-only meals dozens of times. My mum has even made me vegan dishes to take home on dozens of occasions. They almost all acknowledge the horror of factory farming. Veganism is just firmly bucketed a “yanni thing”.
I think that most beings[1] who I’ve become very close with (I estimate this reference class is ≤10) have gone vegan in a way that feels like it was because of our discussions, though it’s possible they would have eventually otherwise, or that our discussions were just the last straw.
In retrospect, I wonder if I had first emanated a framing of taking morality seriously (such that “X is bad” → “I will {not do X} or try to stop X”). I think I also tended to become that close with beings who do already, and who are more intelligent/willing to reflect than what seems normal.
[GIF] A feature I’d love on the forum: while posts are read back to you, the part of the text that is being read is highlighted. This exists on Naturalreaders.com and would love to see it here (great for people who have wandering minds like me)
I agree with you, and so does our issue tracker. Sadly, it does seem a bit hard. Tagging @peterhartree as a person who might be able to tell me that it’s less hard than I think.
For what it’s worth, I would find the first part of the issue (i.e. making the player “floating” or “sticky”) already quite useful, and it seems much easier to implement.
I think acting on the margins is still very underrated. For e.g. I think 5x the amount of advocacy for a Pause on capabilities development of frontier AI models would be great. I also think in 12 months time it would be fine for me to reevaluate this take and say something like ‘ok that’s enough Pause advocacy’.
Basically, you shouldn’t feel ‘locked in’ to any view. And if you’re starting to feel like you’re part of a tribe, then that could be a bad sign you’ve been psychographically locked in.
I’m pretty confident that Marketing is in the top 1-3 skill bases for aspiring Community / Movement Builders.
When I say Marketing, I mean it in the broad sense it used to mean. In recent years “Marketing” = “Advertising”, but I use the classic Four P’s of Marketing to describe it.
The best places to get such a skill base is at FMCG / mass marketing organisations such as the below. Second best would be consulting firms (McKinsey & Company):
Procter & Gamble (P&G)
Unilever
Coca-Cola
Amazon
1. Product - What you’re selling (goods or services) - Features and benefits - Quality, design, packaging - Brand name and reputation - Customer service and support
Yesterday Greg Sadler and I met with the President of the Australian Association of Voice Actors. Like us, they’ve been lobbying for more and better AI regulation from government. I was surprised how much overlap we had in concerns and potential solutions: 1. Transparency and explainability of AI model data use (concern)
2. Importance of interpretability (solution)
3. Mis/dis information from deepfakes (concern)
4. Lack of liability for the creators of AI if any harms eventuate (concern + solution)
5. Unemployment without safety nets for Australians (concern)
6. Rate of capabilities development (concern)
They may even support the creation of an AI Safety Institute in Australia. Don’t underestimate who could be allies moving forward!
Flaming hot take: I wonder if some EAs suffer from Scope Oversensitivity—essentially the inverse of the identifiable victim effect. Take the animal welfare vs global health debate: are we sometimes biased by the sheer magnitude of animal suffering numbers, rather than other relevant factors? Just as the identifiable victim effect leads people to overweight individual stories, maybe we’re overweighting astronomical numbers.
EAs pride themselves on scope sensitivity to combat emotional biases, but taken to an extreme, could this create its own bias? Are we sometimes too seduced by bigger numbers = bigger problem? The meta-principle might be that any framework, even one designed to correct cognitive biases, needs wisdom and balance to avoid becoming its own kind of distortion.
Scope insensitivity has some empirical backing—e.g. the helping birds study—and some theorised mechanisms of action, e.g. people lacking intuitive understanding of large numbers.
Scope oversensitivity seems possible in theory, but I can’t think of any similar empirical or theoretical reasons to think it’s actually happening.
To the extent that you disagree, it’s not clear to me whether it’s because you and I disagree on how EAs weight things like animal suffering, or whether we disagree on how it ought to be weighted. Are you intending to cast doubt on the idea that a problem that is 100x as large is (all else equal) 100x more important, or are you intending to suggest that EAs treat it as more than 100x as important?
Do you have any particular reason to believe that EAs overweight large problems?
Take the animal welfare vs global health debate: are we sometimes biased by the sheer magnitude of animal suffering numbers, rather than other relevant factors?
EAs donate much more to global health than to animal welfare. Do you think the ratio should be even higher still?
The general public wants frontier AI models regulated and there doesn’t seem to be grassroots focussed orgs attempting to capture and funnel this energy into influencing politicians. E.g. via this kind of activity. This seems like massively low hanging fruit. An example of an organisation that does this (but for GH&W) is Results Australia. Someone should set up such an org.
My impression is that PauseAI focusses more on media engagement + protests, which I consider a good but separate thing. Results Australia, as an example, focuses (almost exclusively) on having concerned citizens interacting directly with politicians. Maybe it would be a good thing for orgs to focus on separate things (e.g. for reasons of perception + specialisation). I lean in this direction but remain pretty unsure.
Founder of PauseAI here. I know our protests are the most visible, but they are actually a small portion of what we do. People always talk about the protests, but I think we actually had most results through invisible volunteer lobbying. Personally, I’ve spent way more time sending emails to politicians and journalists, meeting them and informing them of the issues. I wrote an Email Builder to get volunteers to write to their representatives, gave multiple workshops on doing so, and have seen many people (including you of course!) take action and ask for feedback in the discord.
I think combining both protesting and volunteer lobbying in one org is very powerful. It’s not a new idea of course—orgs like GreenPeace have been using this strategy for decades. The protests create visibility and attention, and the lobbying gives the important people the right background information. The protests encourage more volunteers to join and help out, so we get more volunteer lobbyists! In my experience the protests also help with getting an in with politicians—it creates a recognizable brand that shows you represent a larger group.
Is anyone in the AI Governance-Comms space working on what public outreach should look like if lots of jobs start getting automated in < 3 years?
I point to Travel Agents a lot not to pick on them, but because they’re salient and there are lots of them. I think there is a reasonable chance in 3 years that industry loses 50% of its workers (3 million globally).
People are going to start freaking out about this. Which means we’re in “December 2019” all over again, and we all remember how bad Government Comms were during COVID.
Now is the time to start working on the messaging!
I believe we should worry about the pathology of people like Sam Altman, mostly because of what they’re trying to achieve, without the consent of all governments and their respective populace. I found a tweet that summarises my own view better than I can. The pursuit of AGI isn’t normal or acceptable and we should call it out as such.
Remember: EA institutions actively push talented people into the companies making the world changing tech the public have said THEY DONT WANT. This is where the next big EA PR crisis will come from (50%). Except this time it won’t just be the tech bubble.
Is this about the safety teams at capabilities labs?
If so, I consider it a non-obvious issue, whether pushing a talented people into an AI safety role at, e.g., DeepMind is a bad thing. If you think that is a bad thing, consider providing a more detailed argument, and writing a top-level post explaining your view.
If, instead, this is about EA institutions pushing people into capabilities roles, consider naming these concrete examples. As an example, 80k has a job advertising a role as a prompt engineer at Scale AI. That does not seem to be a very safety-focused role, and it is not clear how 80k wants to help prevent human extinction with that job ad.
Thanks for the comment :) I think expecting this detail is normal from someone on the EA Forum. If you think that’s the way public relations are going to work, which is my main point, then we’re disagreeing.
So I did a quick check today—I’ve sent 19 emails to politicians about AI Safety / x-risk and received 4 meetings. They’ve all had a really good vibe, and I’ve managed to get each of them to commit to something small (e.g. email XYZ person about setting up an AI Safety Institute). I’m pretty happy with the hit rate (4/19). I might do another forum quick take once I’ve sent 50.
I expect (~ 75%) that the decision to “funnel” EAs into jobs at AI labs will become a contentious community issue in the next year. I think that over time more people will think it is a bad idea. This may have PR and funding consequences too.
I expect (~ 75%) that the decision to “funnel” EAs into jobs at AI labs will become a contentious community issue in the next year.
My understanding is that this has been a contentious issue for many years already.
80,000 hours wrote a response to this last year, Scott Alexander had written about it in 2022, and Anthropic split from OpenAI in 2021. Do you mean that you expect this to become significantly more salient in the next year?
Yeah my prediction lacked specificity. I expect it to become quite heated. I’m imagining 10+ posts on the topic next year with a total of 100+ comments. That’s just on the forum.
I’d probably bet against this happening FWIW. Maybe a Manifold market?
Also, 100+ comments on the forum might not mean it’s necessarily “heated”—a back and forth between two commenters can quickly increase the tally on any subject so that part might also need specifying further.
I spent about 30 seconds thinking about how to quantify my prediction. I’m trying to point at something vague in a concrete way but failing. This also means that I don’t think it is worth my time making it more concrete. The initial post was more of the “I am pretty confident this will be a community issue, just a heads up”.
1. 80k don’t claim to only advertise impactful jobs
They also advertise jobs that help build career impact, and they’re not against posting jobs that cause harm (and it’s often/always not clear which is which). See more in this post.
They sometimes add features like marking “recommended orgs” (which I endorse!), and sometimes remove those features ( 😿 ).
2. 80k’s career guide about working at AI labs doesn’t dive into “which lab”
We’re really not sure. It seems like OpenAI, Google DeepMind, and Anthropic are currently taking existential risk more seriously than other labs.
I think [link to comment] the “which lab” question is really important, and I’d encourage 80k to either be opinionated about it, or at least help people make up their mind somehow, not just leave people hanging on “which lab” while also often recommending people go work at AI labs, and also mentioning that often that work is net-negative and recommending reducing the risk by not working “in certain positions unless you feel awesome about the lab”.
[I have longer thoughts on how they could do this, but my main point is that it’s (imo) an important hole in their recommendation that might be hidden from many readers]
3. Counterfactual / With great power comes great responsibility
If 80k wouldn’t do all this, should we assume there would be no job board and no guides?
I claim that something like a job board has critical mass: Candidates know the best orgs are there, and orgs know the best candidates are there.
Once there’s a job board with critical mass, it’s not trivial to “compete” with it.
But EAs love opening job boards. A few new EA job boards pop up every year. I do think there would be an alternative. And so the question seems to me to be—how well are 80k using their critical mass?
4. What results did 80k’s work actually cause?
First of all: I don’t actually know, and if someone from 80k would respond, that would be way better than my guess.
Still, here’s my guess, which I think would be better than just responding to the poll:
Lots of engineers who care about AI Safety but don’t have a deep understanding of it (and not much interest in spending months to learn) - go work at AI labs.
Is this positive, because now people “care” in the labs, or negative because the labs have a much easier time hiring people who basically go and do their job? This is seriously a hard question, but I’d guess “negative” (and I think 80k would agree but I’m not sure)
I wouldn’t be surprised if 80k are directly responsible for a few very important hires.
For example, I think the CISO (head of security) of Anthropic used to run the security of Chrome. I’m VERY happy Anthropic hired such a person, I think infosec is really important for AI labs, and I wouldn’t be surprised if 80k had something to do with this, and if not this—then maybe some other similar role.
I think this is very positive, and maybe more important than all the rest.
[I need to go, my comment seems incomplete but I hope it’s still somewhat helpful so posting. I’m still not sure how to vote!]
And there’s always the other option that I (unpopularly) believe in—that better publicly available AI capabilities are necessary for meaningful safety research, thus AI labs have contributed positively to the field.
NotebookLM is basically magic. Just take whatever Forum post you can’t be bothered reading but know you should and use NotebookLM to convert it into a podcast.
It seems reasonable that in 6 − 12 months there will be a button inside each Forum post that converts said post into a podcast (i.e. you won’t need to visit NotebookLM to do it).
It breaks my heart when I see eulogy posts on the forum. And while I greatly appreciate people going to the effort of writing them (while presumably experiencing grief), it still doesn’t feel like enough. We’re talking about people that dedicated their lives to doing good, and all they get is a post. I don’t have a suggestion to address this ‘problem’, and some may even feel that a post is enough, but I don’t. Maybe there is no good answer and death just sucks. I dunno.
I think https://www.wakingup.com/ should be considered for effective organisation status. It donates 10% of revenue to the most effective causes and I think reaching nondual states of awakening could be one of the most effective ways for people in rich countries to improve their wellbeing.
A piece of career advice I’ve given a few times recently to people in AI Safety, which I thought worth repeating here, is that AI Safety is so nascent a field that the following strategy could be worth pursuing:
1. Write your own job description (whatever it is that you’re good at / brings joy into your life).
2. Find organisations that you think need thing that job but don’t yet have it. This role should solve a problem they either don’t know they have or haven’t figured out how to solve.
3. Find the key decision maker and email them. Explain the (their) problem as you see it and how having this role could fix their problem. Explain why you’re the person to do it.
I think this might work better for mid-career people, but, if you’re just looking to skill up and don’t mind volunteering, you could adapt this approach no matter what stage you’re at in your career.
I often see EA orgs looking to hire people to do Z, where applicants don’t necessarily need experience doing Z to be successful.
E.g. instead of saying “must have minimum 4 years as an Operations Manager”, they say “we are hiring for an Operations Manager. But you don’t need experience as an Operations Manager, as long as you have competencies/skills (insert list) A—F.”
This reminds of when Bloomberg spent over $10M training a GPT-3.5 class AI on their own financial data, only to find that GPT-4 beat it on almost all finance tasks.
They thought proprietary data would perform better, but it turns out the largest models won (at the time at least).
I worry that saying “4 years of experience as Operations Manager” is less important than the “competencies A—F” we’ve identified, makes the same mistake.
You’re gambling on having found the secret sauce. But as usual the more targeted the intervention, the more likely it is to miss.
If someone claims to have competencies A-F required to be an operation manager, they should have some way of proving that they actually have competencies A-F. A great way for them to prove this is to have 4 years of experience as Operations Manager. In fact, I struggle to think of a better way to prove this, and obviously a person with experience is preferable to someone with no experience for all kind of reasons (they have encountered all the little errors, etc). As a bonus, somebody outside of your personal bubble has trusted this person.
EA should be relying on more subject matter experience, not less.
To me, your examples at the micro level don’t make the case that you can’t solve a problem at the level it’s created. I’m agnostic as to whether CBT or meta-cognitive therapy is better for intrusive thoughts, but lots of people like CBT; and as for ‘doing the dishes’, in my household we did solve the problem of conflicts around chores by making a spreadsheet. And to the extent that working on communication styles is helpful, that’s because people (I’d claim) have a problem at the level of communication styles.
This is an extremely “EA” request from me but I feel like we need a word for people (i.e. me) who are Vegans but will eat animal products if they’re about to be thrown out. OpportuVegan? UtilaVegan?
If you predictably do this, you raise the odds that people around you will cook some/ buy some extra food so that it will be “thrown out”, or offer you food they haven’t quite finished (and that they’ll replace with a snack later. So I’d recommend going with “Vegan” as your label, for practical as well as signalling reasons.
I am 90% sure that most AI Safety talent aren’t thinking hard enough about what Neglectedness. The industry is so nascent that you could look at 10 analogous industries, see what processes or institutions are valuable and missing and build an organisation around the highest impact one.
The highest impact job ≠ the highest impact opportunity for you!
you could look at 10 analogous industries, see what processes or institutions are valuable
I feel like you’re making this sound simple when I’d expect questions like this to involve quite a bit of work, and potentially skills and expertise that I wouldn’t expect people at the start of their careers to have yet.
Do you have any specific ideas for something that seems obviously missing to you?
AI Safety (in the broadest possible sense, i.e. including ethics & bias) is going be taken very seriously soon by Government decision makers in many countries. But without high quality talent staying in their home countries (i.e. not moving to UK or US), there is a reasonable chance that x/c-risk won’t be considered problems worth trying to solve. X/c-risk sympathisers need seats at the table. IMO local AIS movement builders should be thinking hard about how to either keep talent local (if they’re experiencing brain drain) OR increase the amount of local talent coming into x/c-risk Safety, such that outgoing talent leakage isn’t a problem.
What types of influence do you think governments from small, low influence countries will be able to have?
For example, the NZ government—aren’t they price-takers when it comes to AI regulation? If you’re not a significant player, don’t have significant resources to commit to the problem, and don’t have any national GenAI companies—how will they influence the development trajectory of AI?
That’s one way, but countries’s (and other body’s) concerns are made up of both their citizens/participants concerns and other people’s concerns. one is valued differently from the other, but there’s other ways of keeping relevant actors away from not thinking its a problem. (namely, utilizing the UN.) you make a good point though.
I beta tested a new movement building format last night: online networking. It seems to have legs.
V quick theory of change: > problem it solves: not enough people in AIS across Australia (especially) and New Zealand are meeting each other (this is bad for the movement and people’s impact). > we need to brute force serendipity to create collabs. > this initiative has v low cost
quantitativeresults: > I purposefully didn’t market it hard because it was a beta. I literally got more people that I hoped for > 22 RSVPs and 18 attendees > this says to me I could easily get 40+ > average score for below question was 7.27, which is very good for a beta test
I used Zoom, which was extremely clunky. These results suggest to me I should; > invest in software designed for this use case, not zoom
> segment by career stream (governance vs technical) and/or experience (beginner vs advanced) > run it every second month
I have heaps of qualitative feedback from participants but don’t have time to share it here.
Two jobs in AI Safety Advocacy that AFAICT don’t exist, but should and probably will very soon. Will EAs be the first to create them though? There is a strong first mover advantage waiting for someone −
1. Volunteer Coordinator—there will soon be a groundswell from the general population wanting to have a positive impact in AI. Most won’t know how to. A volunteer manager will help capture and direct their efforts positively, for example, by having them write emails to politicians
2. Partnerships Manager—the President of the Voice Actors guild reached out to me recently. We had a very surprising number of cross over in concerns and potential solutions. Voice Actors are the canary in the coal mine. More unions (etc) will follow very shortly. I imagine within 1 year there will be a formalised group of these different orgs advocating together.
I think the average EA worries too much about negative PR related to EA. I think this is a shame because EA didn’t get to where it is now because it concerned itself with PR. It got here through good-old-fashioned hard work (and good thinking ofc).
Two examples:
1. FTX.
2. OpenAI board drama.
On the whole, I think there was way too much time spent thinking and talking about these massive outliers and it would have been better if 95%+ of EAs put their head down and did what they do best—get back to work.
I think it is good to discuss and take action on the issues at a concrete level, but not to talk about how this affects “what people think of EA”. EA overindexing on conscientiousness can have negative consequences!
“Pens Down” to mean ‘Artificial Super Intelligence in my opinion is close enough that it no longer makes sense to work on whatever else I’m currently working on, because we’re about to undergo radical amounts of change very soon/quickly’.
For me, it is probably when we have something as powerful as GPT-4 except it is agentic and costs less than $100 / month. So, that looks like a digital personal assistant that can execute an instruction like “have a TV delivered for me by X date, under Y price and organise installation and wall mounting.”
This is obviously a question mainly for people who don’t work full time on AI Safety.
I don’t know if this can be answered in full-generality.
I suppose it comes down to things like: • Financial runway/back-up plans in case your prediction is wrong • Importance of what you’re doing now • Potential for impact in AI safety
Thanks Yanni, I think a lot of people have been concerned about this kind of thing.
I would be surprised if 80,000 hours isn’t already tracking this or something like it—perhaps try reaching out to them directly, you might get a better response that way
Three million people are employed by the travel (agent) industry worldwide. I am struggling to see how we don’t lose 80%+ of those jobs to AI Agents in 3 years (this is ofc just one example). This is going to be an extremely painful process for a lot of people.
I think the greater threat to travel agents was the rise of self-planning travel internet sites like skyscanner and booking.com which make them mostly unnecessary. If travel agents have survived that, I don’t see how they wouldn’t survive LLM’s: presumably one of the reasons they survived is the preference for talking in person to a human being.
Yep, I know people who even opt against self-checkout because they enjoy things like a sense of community! I don’t claim AI isn’t going to greatly disrupt a number of jobs, but service jobs in particular seem like some of the most resistant, especially if people get wealthier from AI and have more money to spend on niceties.
I’m not sure it going to feel like a zoom call with a colleague any time soon. That’s a pretty high bar IMO that we ain’t anywhere near yet. Many steps aren’t yet there which include
(I’m probably misrepresenting a couple of these due to lack of expertise but something like...)
1) Video quality (especially being rendered real time) 2) Almost insta-replying 3) Facial warmth and expressions 4) LLM sounding exactly like a real person (this one might be closest) 5) Reduction in processing power required for this to be a norm. Having thousands of these going simultaneously is going to need a lot. (again less important)
I would bet against Avatars being this high fidelity in 3 years a in common use because I think LLM progress is tailing off and there are multiple problems to be solved to get there—but maybe I’m a troglodyte...
That would not contribute to my friends’ sense of community or trust in other people (or, for that matter, the confidence that if a real service person gives you a bad experience or advice, that you have a path to escalate it)
I always get downvoted when I suggest that (1) if you have short timelines and (2) you decide to work at a Lab then (3) people might hate you soon and your employment prospects could be damaged.
What is something obvious I’m missing here?
One thing I won’t find convincing is someone pointing at the Finance industry post GFC as a comparison.
I believe the scale of unemployment could be much higher. E.g. 5% ->15% unemployment in 3 years.
If you have short timelines, I feel like your employment prospects in a fairly specific hypothetical should be fairly low on your priority list compared to ensuring AI goes well? Further, if you have short timelines, knowing about AI seems likely to be highly employable and lab experience will be really useful (until AI can do it better than us...)
I can buy “you’ll be really unpopular by affiliation and this will be bad for your personal life” as a more plausible argument
“If you have short timelines, I feel like your employment prospects in a fairly specific hypothetical should be fairly low on your priority list compared to ensuring AI goes well?”
- Possibly! I’m mostly suggesting that people should be aware of this possibility. What they do with that info is obvs up to them
”if you have short timelines, knowing about AI seems likely to be highly employable and lab experience will be really useful”
- The people I’m pointing to here (e.g. you) will be employable and useful. Maybe we’re talking passed each other a bit?
The people I’m pointing to here (e.g. you) will be unemployable or unuseful. Maybe we’re talking passed each other a bit?
Why do you believe this? Being an expert in AI is currently very employable, and seems all the more so if AI is automating substantial parts of the economy
I’m very confused. Your top level post says “your employment prospects could be damaged.” I am explaining why I think that the skillset of lab employees will be way more employable. Is your argument that the PR will outweigh that, even though it’s a major factor?
Financial workers’ careers are not directly affected very much by populist prejudice because 1) they provide a useful service and ultimately people pay for services and 2) they are employed by financial firms who do not share that prejudice. Likewise in short timeline worlds AI companies are probably providing a lot of services to the world, allowing them to pay high compensation, and AI engineers would have little reason to want to work elsewhere. So even if they are broadly disliked (which I doubt) they could still have very lucrative careers.
Obviously its a different story if antipathy towards the financial sector gets large enough to lead to a communist takeover.
I believe the scale of unemployment could be much higher. E.g. 5% ->15% unemployment in 3 years.
During the financial crisis U-3 rose from 5% to 10% over about 2 years, suggesting on this metric at least you should think it is at least comparable (2.5%/year vs 3.3%/year).
Just on that last prediction—that’s an extraordinary claim given that the U.S. unemployment rate has only increased 0.7% since ChatGPT was released nearly 2 years ago (harder to compare again GPT-3, which was released in 2020, but you could say the same about the economic impacts of recent LLMs more broadly). If you believe we’re on the precipice of some extraordinary increase, why stop at 15%? Or if we’re not on that precipice, what’s the mechanism for that 10% jump?
I’m pretty confident that a majority of the population will soon have very negative attitudes towards big AI labs. I’m extremely unsure about what impact this will have on the AI Safety and EA communities (because we work with those labs in all sorts of ways). I think this could increase the likelihood of “Ethics” advocates becoming much more popular, but I don’t know if this necessarily increases catastrophic or existential risks.
Anecdotally I feel like people generally have very low opinion of and understanding of the finance sector, and the finance sector mostly doesn’t seem to mind. (There are times when there’s a noisy collision, e.g. Melvin Capital vs. GME, but they’re probably overall pretty rare.)
It’s possible / likely that AI and tech companies targeting mass audiences with their product are more dependent on positive public perception, but I suspect that the effect of being broadly hated is less strong than you’d think.
I doubt the prevalence of AI is making people more positive towards the sector given all the negative publicity over plagarism, job loss, and so on. So I would guess the public already dislikes AI companies (even if they use their products), and this will probably increase.
I’m pretty confident that a majority of the population will soon have very negative attitudes towards big AI labs.
Can you elaborate on what makes you so certain about this? Do you think that the reputation will be more like that of Facebook or that of Big Tobacco? Or will it be totally different?
More thoughts on “80,000 hours should remove OpenAI from the Job Board”
- - - -
A broadly robust heuristic I walk around with in my head is “the more targeted the strategy, the more likely it will miss the target”.
Its origin is from when I worked in advertising agencies, where TV ads for car brands all looked the same because they needed to reach millions of people, while email campaigns were highly personalised.
The thing is, if you get hit with a tv ad for a car, worst case scenario it can’t have a strongly negative effect because of its generic content (dude looks cool while driving his 4WD through rocky terrain). But when an email that attempts to be personalised misses the mark, you feel it (Hi {First Name : Last Name}!).
Anyway, effective altruism loves targeting. It is virtually built from the ground up on it (RCTs and all that, aversion to PR etc).
Anyway, I am thinking about this in relation to the recent “80,000 hours should remove OpenAI from the Job Board” post.
80k is applying an extremely targeted strategy by recommending people work in some jobs at the workplaces creating the possible death machines (caveat 1, caveat 2, caveat 3, caveatn).
There are so many failure modes involved with those caveats that makes it virtually impossible to not go catastrophically poorly in at least a few instances and take the mental health of a bunch of young and well-meaning EAs with it.
A judgement I’m attached to is that a person is either extremely confused or callous if they work in capabilities at a big lab. Is there some nuance I’m missing here?
Some of them have low p(doom from AI) and aren’t longtermists, which justifies the high-level decision of working on an existentially dangerous technology with sufficiently large benefits.
I do think their actual specific actions are not commensurate with the level of importance or moral seriousness that they claim to attach to their work, given those stated motivations.
Thanks for the comment Linch! Just to spell this out:
“Some of them have low p(doom from AI) and aren’t longtermists, which justifies the high-level decision of working on an existentially dangerous technology with sufficiently large benefits. ”
I would consider this acceptable if their p(doom) was ~ < 0.01%
I find this pretty incredulous TBH
“I do think their actual specific actions are not commensurate with the level of importance or moral seriousness that they claim to attach to their work, given those stated motivations.”
I’m a bit confused by this part. Are you saying you believe the importance / seriousness the person claims their work has is not reflected in the actions they actually take? In what way are you saying they do this?
I would consider this acceptable if their p(doom) was ~ < 0.01%
I find this pretty incredulous TBH
I dunno man, at least some people think that AGI is the best path to curing cancer. That seems like a big deal! If you aren’t a longtermist at all, speeding up the cure for cancer by a year is probably worth quite a bit of x-risk.
Are you saying you believe the importance / seriousness the person claims their work has is not reflected in the actions they actually take?
Yes.
In what way are you saying they do this?
Lab people shitpost, lab people take competition extremely seriously (if your actual objective is “curing cancer,” it seems a bit discordant to be that worried that Silicon Valley Startup #2 will cure cancer before you), people take infosec not at all seriously, the bizarre cultish behavior after the Sam Altman firing, and so forth.
In my experience, they’re mostly just impulsive and have already written their bottom line (“I want to work on AI projects that look cool to me”), and after that they come up with excuses to justify this to themselves.
Sorry, I will explain further for clarity. I am Buddhist and in that philosophy the phrase “confused” has an esoteric usage. E.g. “that person suffers because they’re confused about the nature of reality”. I think this lab person is confused about how much suffering they’re going to create for themselves and others (i.e. it hasn’t been seen or internalised).
Yes. I think most people working on capabilities at leading labs are confused or callous (or something similar, like greedy or delusional), but definitely not all. And personally, I very much hope there are many safety-concerned people working on capabilities at big labs, and am concerned about the most safety-concerned people feeling the most pressure to leave, leading to evaporative cooling.
Reasons to work on capabilities at a large lab:
To build career capital of the kind that will allow you to have a positive impact later. E.g. to be offered relevant positions in government
To positively influence the culture of capabilities teams or leadership at labs.
To be willing and able to whistleblow bad situations (e.g. seeing emerging dangerous capabilities in new models, the non-disparagement stuff).
[maybe] to earn to give (especially if you don’t think you’re contributing to core capabilities)
To be clear, I expect achieving the above to be infeasible for most people, and it’s important for people to not delude themselves into thinking they’re having a positive impact to keep enjoying a lucrative, exciting job. But I definitely think there are people for whom the above is feasible and extremely important.
Another way to phrase the question is “is it good for all safety-concerned people to shun capabilities teams, given (as seems to be the case) that those teams will continue to exist and make progress by default?” And for me the strong answer is “yes”. Which is totally consistent with wanting labs to pause and thinking that just contributing to capabilities (on frontier models) in expectation is extremely destructive.
Thanks so much for your thoughtful comment! I appreciate someone engaging with me on this rather than just disagree ticking. Some thoughts:
1. Build Career Capital for Later Impact:
I think this depends somewhat on what your AGI timelines are. If they’re short, you’ve wasted your time and possibly had a net-negative effect.
I think there is a massive risk of people entering an AGI lab for career capital building reasons, and then post rationalising their decision to stay. Action changes attitudes faster than attitude changes actions after all.
2. Influence the Culture:
Multiple board members attempted this at OpenAI got fired, what chance does a single lower level employee?
I have tried changing corporate culture at multiple organisations and it is somewhere between extremely hard and impossible (more on that below).
3. Be Prepared to Whistle-Blow:
I was a Whistle-Blower during my time at News Corp. It was extremely difficult. I simply do not expect a meaningful number of people to be able to do this. You have to be willing to turn your life upside down.
This can be terrible for a person’s mental health. We shouldn’t be vary careful openly recommending this as a reason to stay at a Lab.
As mentioned above, I expect a greater number of people to turn into converts than would-be Whistle-Blowers
I think if you’re the kind of person who goes into a Lab for Career Capital that makes you less likely to be a Whistle-Blower TBH.
(1) I agree if your timelines are super short, like <2yrs, it’s probably not worth it. I have a bunch of probability mass on longer timelines, though some on really short ones
Re (2), my sense is some employees already have had some of this effect (and many don’t. But some do). I think board members are terrible candidates for changing org culture; they have unrelated full-time jobs, they don’t work from the office, they have different backgrounds, most people don’t have cause to interact with them much. People who are full-time, work together with people all day every day, know the context, etc., seem more likely to be effective at this (and indeed, I think they have been, to some extent in some cases)
Re (3), seems like a bunch of OAI people have blown the whistle on bad behavior already, so the track record is pretty great, and I think them doing that has been super valuable. And 1 whistleblower seems much better than several converts is bad. I agree it can be terrible for mental health for some people, and people should take care of themselves.
Re (4), um, this is the EA Forum, we care about how good the money is. Besides crypto, I don’t think there are many for many of the relevant people to make similar amounts of money on similar timeframes. Actually I think working at a lab early was an effective way to make money. A bunch of safety-concerned people for example have equity worth several millions to tens of millions, more than I think they could have easily earned elsewhere, and some are now billionaires on paper. And if AI has the transformative impact on the economy we expect, that could be worth way more (and it being worth more is correlated with it being needed more, so extra valuable); we are talking about the most valuable/powerful industry the world has ever known here, hard to beat that for making money. I don’t think that makes it okay to lead large AI labs, but for joining early, especially doing some capabilities work that doesn’t push the most risky capabilities along much, I don’t think it’s obvious.
I agree that there are various risks related to staying too long, rationalizing, being greedy, etc., and in most cases I wouldn’t advice a safety-concerned person to do capabilities. But I think you’re being substantially too intense about the risk of speeding up AI relative to the benefits of seeing what’s happening on the inside, which seem like they’ve already been very substantial
The Greg / Vasco bet reminded me of something: I went to buy a ceiling fan with a light in it recently. There was one on sale that happened to also tick all my boxes, joy! But the salesperson warned me “the light in this fan can’t be replaced and only has 10,000 hours in it. After that you’ll need a new fan. So you might not want to buy this one.” I chuckled internally and bought two of them, one for each room.
Has anyone seen an analysis that takes seriously the idea that people should eat some fruits, vegetables and legumes over others based on how much animal suffering they each cause?
I.e. don’t eat X fruit, eat Y one instead, because X fruit is [e.g.] harvested in Z way, which kills more [insert plausibly sentient creature].
Even though I’ve been in the AI Safety space for ~ 2 years, I can’t shake the feeling that every living thing dying painlessly in its sleep overnight (due to AI killing us) isn’t as bad as (i.e. is ‘better’ than) hundreds of millions of people living in poverty and/or hundreds of billions of animals being tortured.
This makes me suspicious of my motivations. I think I do the work partly because I kinda feel the loss of future generations, but mainly because AI Safety still feels so neglected (and my counter factual impact here is larger).
I don’t think s-risks play much of a role in this, although they have in the past (here, I define disempowerment due to AGI, or authoritarian use of AGI as s-risks).
Thanks for sharing. I suspect most of the hundreds of millions of people living in poverty would disagree with you, though, and would prefer not to painlessly die in their sleep tonight.
I don’t think he’s talking past you. His point seems that the vast majority of the hundreds of millions of people living in poverty both have net positive lives, and don’t want to die.
Even with a purely hedonistic outlook, it wouldn’t be better for their lives to end.
Unless you are not talking about the present, but a future far worse than today’s situation?
I’m saying that on some level it feels worse to me that 700 million people suffer in poverty than every single person dying painlessly in their sleep. Or that billions of animals are in torture factories. It sounds like I’m misunderstanding Jason’s point?
I would contend they are not “suffering” in poverty overall, because most of their lives are net positive. There may be many struggles and their lives are a lot harder than ours, but still better than not being alive at all.
I agree with you on the animals in torture factories, because their lives are probably net negative unlike the 700 million in poverty.
If AI actually does manage to kill us (which I doubt), It will not involve everybody dying painlessly in their sleep. That is an assumption of the “FOOM to god AI with no warning” model, which bears no resemblance to reality.
The technology to kill everyone on earth in their sleep instantaneously does not exist now, and will not exist in the near-future, even if AGI is invented. Killing everyone in their sleep is orders of magnitude more difficult than killing everyone awake, so why on earth would that be the default scenario?
I think you have a point with animals, but I don’t think the balance of human experience means that non-existence would be better than the status quo.
Will talks about this quite a lot in ch. 9 of WWOTF (“Will the future be good or bad?”). He writes:
If we assume, following the small UK survey, that the neutral point on a life satisfaction scale is between 1 and 2, then 5 to 10 percent of the global population have lives of negative wellbeing. In the World Values Survey, 17 percent of respondents classed themselves as unhappy. In the smaller skipping study of people in rich countries, 12 percent of people had days where their bad experiences outweighed the good. And in the study that I commissioned, fewer than 10 percent of people in both the United States and India said they wished they had never been born, and a little over 10 percent said that their lives contained more suffering than happiness.
So, I would guess that on either preference-satisfactionism or hedonism, most people have lives with positive wellbeing. If I were given the option, on my deathbed, to be reincarnated as a randomly selected person alive today, I would choose to do so.
And, of course, for people at least, things are getting better over time. I think animal suffering complicates this a lot.
Imagine running 100 simulations of humanity’s story. In every single one, the same pattern emerges: The moment we choose agriculture over hunting and gathering, we unknowingly start a countdown to our own extinction through AGI. If this were true, I think it suggests that our best chance at long-term survival is to stay as hunter-gatherers—and that what we call ‘progress’ is actually a kind of cosmic trap.
What makes the agriculture transition stand out as the “point of no return”? Agriculture was independently invented several times, so I’d expect the hunter-gatherer → agriculture transition to be more convergent than agriculture → AGI.
I expect that over the next couple of years GenAI products will continue to improve, but concerns about AI risk won’t. For some reason we can’t “feel” the improvements. Then (e.g. around 2026) we will have pretty capable digital agents and there will be a surge of concern (similar to when ChatGPT was released, maybe bigger), and then possibly another perceived (but not real) plateau.
The concerns in this case seem to be proportional to the actual real world impact: The jump from chatgpt and midjourney not existing to being available to millions was extremely noticeable: suddenly the average person could chat with a realistic bot and use it to cheat on their homework, or configure up a realistic looking image from a simple text prompt.
In contrast for the average person, not a lot has changed from 2022 to now. The chatbots give out less errors and the images are a lot better, but they aren’t able to accomplish significantly more actual tasks than they were in 2022. Companies are shoving AI tools into everything, but people are mostly ignoring them.
We seem to be boiling the frog, however I’m optimistic (perhaps naively) that GPT voice mode may wake some people up. “Just a chatbot” doesn’t carry quite the same weight when it’s actually speaking to you.
I really like this ad strategy from BlueDot. 5 stars.
90% of the message is still x / c-risk focussed, but by including “discrimination” and “loss of social connection” they’re clearly trying to either (or both);
create a big tent
nudge people that are in “ethics” but sympathetic to x / c-risk into x / c-risk
This feels like misleading advertising to me, particularly “Election interference” and “Loss of social connection”. Unless bluedot is doing some very different curriculum now and have a very different understanding of what alignment is. These bait-and-switch tactics might not be a good way to create a “big tent”.
Not sure if this type of concern has reached the meta yet, but if someone approached me asking for career advice, tossing up whether to apply for a job at a big AI lab, I would let them know that it could negatively affect their career prospects down the track because so many people now perceive such as a move as either morally wrong or just plain wrong-headed. And those perceptions might only increase over time. I am not making a claim here beyond this should be a career advice consideration.
Attention: AI Safety Enthusiasts in Wellington (and New Zealand)
I’m pleased to announce the launch of a brand new Facebook group dedicated to AI Safety in Wellington: AI Safety Wellington (AISW). This is your local hub for connecting with others passionate about ensuring a safe and beneficial future with artificial intelligence / reducing x-risk.To kick things off, we’re hosting a super casual meetup where you can:
Meet & Learn: Connect with Wellington’s AI Safety community.
Chat & Collaborate: Discuss career paths, upcoming events, and training opportunities.
Share & Explore: Whether you’re new to AI Safety or a seasoned expert, come learn from each other’s perspectives.
Expand the Network: Feel free to bring a friend who’s curious about AI Safety!
Something I’m confused about: what is the threshold that needs meeting for the majority of people in the EA community to say something like “it would be better if EAs didn’t work at OpenAI”?
Imagining the following hypothetical scenarios over 2024⁄25, I can’t predict confidently whether they’d individually cause that response within EA?
Ten-fifteen more OpenAI staff quit for varied and unclear reasons. No public info is gained outside of rumours
There is another board shakeup because senior leaders seem worried about Altman. Altman stays on
Superalignment team is disbanded
OpenAI doesn’t let UK or US AISI’s safety test GPT5/6 before release
There are strong rumours they’ve achieved weakly general AGI internally at end of 2025
The catchphrase I walk around with in my head regarding the optimal strategy for AI Safety is something like: Creating Superintelligent Artificial Agents* (SAA) without a worldwide referendum is ethically unjustifiable. Until a consensus is reached on whether to bring into existence such technology, a global moratorium is required (*we already have AGI).
I recently discovered the idea of driving all blames into oneself, which immediately resonated with me. It is relatively hardcore; the kind of thing that would turn David Goggins into a Buddhist.
Gemini did a good job of summarising it:
This quote by Pema Chödron, a renowned Buddhist teacher, represents a core principle in some Buddhist traditions, particularly within Tibetan Buddhism. It’s called “taking full responsibility” or “taking self-blame” and can be a bit challenging to understand at first. Here’s a breakdown:
What it Doesn’t Mean:
Self-Flagellation: This practice isn’t about beating yourself up or dwelling on guilt.
Ignoring External Factors: It doesn’t deny the role of external circumstances in a situation.
What it Does Mean:
Owning Your Reaction: It’s about acknowledging how a situation makes you feel and taking responsibility for your own emotional response.
Shifting Focus: Instead of blaming others or dwelling on what you can’t control, you direct your attention to your own thoughts and reactions.
Breaking Negative Cycles: By understanding your own reactions, you can break free from negative thought patterns and choose a more skillful response.
Analogy:
Imagine a pebble thrown into a still pond. The pebble represents the external situation, and the ripples represent your emotional response. While you can’t control the pebble (the external situation), you can control the ripples (your reaction).
Benefits:
Reduced Suffering: By taking responsibility for your own reactions, you become less dependent on external circumstances for your happiness.
Increased Self-Awareness: It helps you understand your triggers and cultivate a more mindful response to situations.
Greater Personal Growth: By taking responsibility, you empower yourself to learn and grow from experiences.
Here are some additional points to consider:
This practice doesn’t mean excusing bad behavior. You can still hold others accountable while taking responsibility for your own reactions.
It’s a gradual process. Be patient with yourself as you learn to practice this approach.
What are some historical examples of a group (like AI Safety folk) getting something incredibly wrong about an incoming Technology? Bonus question: what led to that group getting it so wrong? Maybe there is something to learn here.
In the 90′s and 2000′s, many people such as Eric Drexler were extremely worried about nanotechnology and viewed it as an existential threat through the “gray goo” scenario. Yudkowsky predicted drexler style nanotech would occur by 2010, using very similar language to what he is currently saying about AGI.
It turned out they were all being absurdly overoptimistic about how soon the technology would arrive, and the whole drexlerite nanotech project flamed out by the end of the 2000′s and has pretty much not progressed since. I think a similar dynamic playing out with AGI is less likely, but still very plausible.
Do you have links to people being very worried about gray goo stuff?
(Also, the post you link to makes this clear, but this was a prediction from when Eliezer was a teenager, or just turned 20, which does not make for a particularly good comparison, IMO)
This is probably a good exercise. I do want to point out a common bias about getting existential risks wrong. If someone was right about doomsday, we would not be here to discuss it. That is a huge survivorship bias. Even catestrophic events which lessen the number of people are going to be systemically underestimated. This phenomenon is the anthropic shadow which is relevant to an analysis like this.
Yeah, Case Studies as Research need to be treated very carefully (i.e. they can still be valuable exercises but the analyser needs to be aware of their weaknesses)
There were many predictions about AI and AGI in the past (maybe mostly last century) that were very wrong. I think I read about it in Superintelligence. A quick Google search shows this article which probably talks about that.
Cultured meat predictions were overly optimistic, although many of those predictions might have been companies hyping up their products to attract investors. There’s also probably a selection bias where the biggest cultured meat optimisits are the ones who become cultured meat experts and make predictions
I have heard rumours that an AI Safety documentary is being made. Separate to this, a good friend of mine is also seriously considering making one, but he isn’t “in” AI Safety. If you know who this first group is and can put me in touch with them, it might be worth getting across each others plans.
One of the seminal texts in marketing science is The Long and The Short of it by Field and Binet. They argue that for maximum effectiveness, marketing should aim to create two distinct work streams and results; immediate sales and longer term sales.
They argue the tactics that go into either are distinct (e.g. short term = create a sense of urgency, long term = match brand associations with category associations).
This feels like a good analogy for AI Safety Advocacy / Governance—keep talking about short term things people can buy now (in an Australian context that is the AI Safety Institute) while maintaining salience on the long term (x-risk, a global Pause, etc).
The question shouldn’t be either/or, rather, its about how much (%) effort should go into either at any point in time.
What would be the pros and cons of adding a semi-hidden-but-permanent Hot Takes section to the Forum? All of my takes are Hot and due to time constraints I would otherwise not post at all. Some would argue that someone like me should not post Hot Takes at all. Anyway, in true lazy fashion here is ChatGPT on the pros and cons:
Pros:
Encourages diverse perspectives and stimulates debate.
Can attract more engagement and interest from users.
Provides a platform for expressing unconventional or controversial ideas.
Fosters a culture of intellectual curiosity and open discourse within the community.
Cons:
May lead to increased polarization and conflict within the community.
Risk of spreading misinformation or poorly researched opinions.
Could divert attention from more rigorous and evidence-based discussions.
Potential for reputational damage if controversial opinions are associated with the forum.
That’s an interesting one—I’m a fan of hot takes myself :D. I think “Quick takes” does the job on these though, even if the posts are a bit longer. I’m not sure we need another section. Maybe a “Hot takes” tab could be added to signify that the thought behind a take isn’t so deep?
Feels kinda obnoxious to write a quick take along the lines of “I’m thinking about writing a post on X, does anyone actually give a sh*t? Otherwise I won’t write it.”
I just wanted to check, since I can’t place my finger on why it feels obnoxious but it certainly does.
How impactful could non-dual meditation be for improving wellbeing?
Are there any EAs out there who have practiced non-dual meditation? Specifically, anything from the Dzogchen or Mahamudra traditions of Tibetan Buddhism?
More secular teachers would be Loch Kelly, Sam Harris, Michael Taft.
This has been a life changing experience for me and I’m wondering whether it could be a blind spot for EA.
I’d also just love to chat with someone else who has experienced non-duality / awakening through this form of meditation :)
I spent some time with Claude this morning trying to figure out why I find it cringe calling myself an EA (I never call myself an EA, even though many in EA would call me an EA).
The reason: calling myself “EA” feels cringe because it’s inherently a movement/community label—it always carries that social identity baggage with it, even when I’m just trying to describe my personal philosophical views.
I am happy to describe myself as a Buddhist or Utilitarian because I don’t think it does those things (at least, not within the broader community context I find myself in—Western, Online, Democratic, Australia, etc).
Prediction: In 6-12 months people are going to start leaving Deepmind and Anthropic for similar sounding reasons to those currently leaving OpenAI (50% likelihood).
> Surface level read of what is happening at OpenAI; employees are uncomfortable with specific safety policies. > Deeper, more transferable, harder to solve problem; no person that is sufficiently well-meaning and close enough to the coal face at Big Labs can ever be reassured they’re doing the right thing continuing to work for a company whose mission is to build AGI.
Basically, this is less about “OpenAI is bad” and more “Making AGI is bad”.
Very quick thoughts on setting time aside for strategy, planning and implementation, since I’m into my 4th week of strategy development and experiencing intrusive thoughts about needing to hurry up on implementation;
I have a 52 week LTFF grant to do movement building in Australia (AI Safety)
I have set aside 4.5 weeks for research (interviews + landscape review + maybe survey) and strategy development (segmentation, targeting, positioning),
Then 1.5 weeks for planning (content, events, educational programs), during which I will get feedback from others on the plan and then iterate it.
This leaves me with 46⁄52 weeks to implement ruthlessly.
In conclusion, 6 weeks on strategy and planning seems about right. 2 weeks would have been too short, 10 weeks would have been too long, this porridge is juuuussttt rightttt.
keen for feedback from people in similar positions.
What would stop you from paying for an LLM? Take an extreme case; Sam Altman turns around tomorrow and says “We’re racing to AGI, I’m not going to worry about Safety at all.”
Would that stop you from throwing him $20 a month?
Help clear something up for me: I am extremely confused (theoretically) how we can simultaneously have:
1. An Artificial Superintelligence
2. It be controlled by humans (therefore creating misuse of concentration of power issues)
My intuition is that once it reaches a particular level of power it will be uncontrollable. Unless people are saying that we can have models 100x more powerful than GPT4 without it having any agency??
I don’t think we need to explicitly alert the reader when we’ve received help from an LLM to write something (even if it wrote the bulk of the post). That’s it, my quickest ever Quick Take.
There have been multiple occasions where I’ve copy and pasted email threads into an LLM and asked it things like:
What is X person saying
What are the cruxes in this conversation?
Summarise this conversation
What are the key takeaways
What views are being missed from this conversation
I really want an email plugin that basically brute forces rationality INTO email conversations.
Tangentially—I wonder if LLMs can reliably convert peoples claims into a % through sentiment analysis? This would be useful for Forecasters I believe (and rationality in general)
Having a nondual Awakening was the second most important thing to happen to me (after my daughters birth). It has led to incredibly low levels of suffering and incredibly high levels of wellbeing. I write this because I think it is still incredibly under-appreciated and attainable for most people (maybe literally anyone).
There are traditions (Dzogchen, Zen, modern nonduality) where this shift in consciousness can be experienced simply by hearing the right combination of words and insights. As our understanding and tools for communicating these insights evolve, including through advances in AI, I believe this transformative experience could become accessible to many more people.
How bad does factory farming need to get before we WANT TO accelerate AI capabilities?
This article suggests that the DAILY “total painfulness of fish suffering is equivalent to the painfulness of one hundred fifty million human deaths” and that most of these deaths are via “slow suffocation, crushing, or live disemboweling”.
Let’s assume the creation of AGI had a >80% probability of leading to two outcomes for non-human animals: extinction or liberation.
At some point over the next 5 years virtually every person reading this will become a Pause advocate, the only question is whether that occurs now, in 2025, 2026, 2027, 2028 or 2029.
Flaming hot take: if you think Digital Sentience should be taken seriously but not Human Awakening / Enlightenment, then EA culture might have its hooks in a bit deep.
I like the “we decided to shut down our charity because it wasn’t very effective post” for the obvious reasons, but I wonder once you control for “is rich / comes from a rich family” and “doesn’t have family that depends on them” how that metric gets affected.
I think it is still good praise the decision in general, but unless I know the backgrounds of the people doing it I can’t heap too much praise on them.
If I’m understanding this concern correctly, it’s along the lines of: “they’re not making a financial sacrifice in shutting down, so it’s less praiseworthy than it otherwise would be”.
Just to clarify, charity founders (at least CE ones) take a pay cut to start their charity—they would earn more if working for other EA organizations as employees, and much more if in tech/finance/consulting/careers that typical of people with oxbridge/ivy/etc education levels. The financial sacrifice was already made when starting the charity, and if anything, quitting is actually better for you financially.
I’m confused about where you’re going with this. Why would the founders’ personal financial situation substantially affect how we viewed their decision?
In the absence of a poll feature, please use the agree/disagree function and the “changed my mind” emoji in this quick take to help me get a sense for EA’s views on a statement:
”Working on capabilities within a leading AI Lab makes someone a bad person”
If you listen to this, and eventually get to a point* where it doesn’t sound like gibberish (i.e. actually internalise), I guarantee your wellbeing will improve at least 10%.
*I imagine it will take on average about 6 months of work for a newbie.
I write this as I believe Awakening is still massively misunderstood by EA
I met Australia’s Assistant Minister for Defence last Friday. I asked him to write an email to the Minister in charge of AI, asking him to establish an AI Safety Institute. He said he would. He also seemed on board with not having fully autonomous AI weaponry.
All because I sent one email asking for a meeting + had said meeting.
Advocacy might be the lowest hanging fruit in AI Safety.
Akash’s Speaking to Congressional staffers about AI risk seems similar:
Like you, Akash just cold-emailed people:
There’s a lot of concrete learnings in that writeup; definitely worth reading I think.
I don’t know how to say this in a way that won’t come off harsh, but I’ve had meetings with > 100 people in the last 6 months in AI Safety and I’ve been surprised how poor the industry standard is for people:
- Planning meetings (writing emails that communicate what the meeting will be about, it’s purpose, my role)
- Turning up to scheduled meetings (like literally, showing up at all)
- Turning up on time
- Turning up prepared (i.e. with informed opinions on what we’re going to discuss)
- Turning up with clear agendas (if they’ve organised the meeting)
- Running the meetings
- Following up after the meeting (clear communication about next steps)
I do not have an explanation for what is going on here but it concerns me about the movement interfacing with government and industry. My background is in industry and these just seem to be things people take for granted?
It is possible I give off a chill vibe, leading to people thinking these things aren’t necessary, and they’d switch gears for more important meetings!
My 2 cents is that I think this is a really important observation (I think it’s helpful) and I would be interested to see what other peoples feedback is on this front.
hypothesis that springs to mind, might or might not be useful for engaging with it productively. might be wrong depending on what class of people you have been having meetings with.
when you select for people working on AI risks, you select for people who are generally less respective of status quo and social norms. you select for the kind of person who will generally do something less, because of the reason that its a social norm. in order for this reference class of person to do a thing, they kind of who through their personal reasoning methods reaches the conclusion “It would be worth the effort for me to change myself to become the kind of person who shows up on time consistently, compared to other things I could be spending my effort on”. they might just figure its a better use of their effort to think about their research all day.
I would guess that most people on this earth don’t show up on time because they reasoned it through that this is a good idea, they do it because it has been drilled into them through social norms, and they valued those social norms higher.
(note: this comment is not intended to be an argument that showing up on time is a waste of time)
It just occurred to me that maybe posting this isn’t helpful. And maybe net negative. If I get enough feedback saying it is then I’ll happily delete.
When AI Safety people are also vegetarians, vegans or reducetarian, I am pleasantly surprised, as this is one (of many possible) signals to me they’re “in it” to prevent harm, rather than because it is interesting.
I am skeptical that exhibiting one of the least cost-effective behaviors for the purpose of reducing suffering would be correlated much with being “in it to prevent harm”. Donating to animal charities, or being open to trade with other people about reducing animal suffering, or other things that are actually effective seem like much better indicator that people are actually in it to prevent harm.
My best guess is that most people who are vegetarians, vegans or reducetarians, and are actually interested in scope-sensitivity, are explicitly doing so for signaling and social coordination reasons (which to be clear, I think has a bunch going for it, but seems like the kind of contingent fact that makes it not that useful as a signal of “being in it to prevent harm”).
Have you actually just talked to people about their motives for reducing their animal product consumption, or read about them? Do you expect them to tell you it’s mostly for signaling or social coordination, or (knowingly or unknowingly) lie that it isn’t?
I’d guess only a minority of EAs go veg mostly for signaling or social coordinaton.
Other reasons I find more plausibly common, even among those who make bigger decisons in scope-sensitive ways:
Finding it cost-effective in an absolute sense, e.g. >20 chickens spared per year, at small personal cost (usually ignoring opportunity costs to help others more). (And then maybe falsely thinking you should go veg to do this, rather than just cut out certain animal products, or finding it easier to just go veg than navigate exceptions.)
Wanting to minimize the harm they personally cause to others, or to avoid participating in overall harmful practices, or deontological or virtue ethics reasons, separately from actively helping others or utilitarian-ish reasons.
Not finding not eating animal products very costly, or not being very sensitive to the costs. It may be easy or take from motivational “budgets” psychologically separate from work or donations.
Finding eating animal products psychologically or time costly, e.g. just feeling guilty and worrying about it, cognitive dissonance.
Seeing so many other EAs veg, and this making it seem very impactful to be veg for animals (whether or not it actually is).
As a way to live out their commitment daily that other animals matter and increase/maintain their concern for them (and/or other vulnerable moral patients easy to ignore).
To get other people around them interested in helping animals.
I’d guess most EAs who go veg or reducetarian do so for reasons like 1, 2, 3, 4 and/or 5. 1, 2, 3 and 4 are why I went vegan, around when I first got into EA. I also used to make arguments like 1, just comparing the harms to animals and your personal costs (after the drowning child thought experiment) when I was an EA organizer a few years ago.
I don’t think the opportunity costs of veg diets are on people’s minds much, so it seems likely that people are ignoring them, either for bad reasons or because they don’t seem important enough to them.
I’m still (bival)vegan, but mostly for reasons like 3, 6, just preferring not to eat animal products anymore, and maybe to meet (my perception of) others’ expectations of me.
I’m not even convinced veganism is better for animals anymore through its supply and demand effects, when you include the effects on wild animals (mostly wild arthropods and fish).
I have had >30 conversations with EA vegetarians and vegans about their reasoning here. The people who thought about it the most seem to usually settle on it for signaling reasons. Maybe this changed over the last few years in EA, but it seemed to be where most people I talked to where at when I had lots of conversations with them in 2018.
I agree that many people say (1), but when you dig into it it seems clear that people incur costs that would be better spent on donations, and so I don’t think it’s good reasoning. As far as I can tell most people who think about it carefully seem to stop thinking its a good reason to be vegan/vegetarian.
I do think the self-signaling and signaling effects are potentially substantial.
I also think (4) is probably the most common reason, and I do think probably captures something important, but it seems like a bad inference that “someone is in it to prevent harm” if (4) is their reason for being vegetarian or vegan.
I’ve thought a lot about this, because I’m serious about budgeting and try to spend as little money as possible to make more room for investments and donations. I also have a stressful job and don’t like to spend time cooking. I did not find it hard to switch to being vegan while keeping my food budget the same and maintaining a high-protein diet.
Pea protein is comparable in cost per gram protein to the cheapest animal products where I live, and it requires no cooking. Canned beans are a bit more expensive but still very cheap and also require no cooking. Grains, ramen and cereal are very cheap sources of calories. Plant milks are more expensive than cow’s milk, but still fit in my very low food budget. Necessary supplements like B12 are very cheap.
On a day-to-day basis, the only real cooking I do is things like pasta, which don’t take much time at all. I often go several weeks without doing any real cooking. I’d bet that I both spend substantially less money on food and substantially less time cooking than the vast majority of omnivores, while eating more protein.
As a vegan it’s also easy to avoid spending money in frivolous ways, like on expensive ready-to-eat snacks and e.g. DoorDash orders.
I haven’t had any health effects either, or any differences in how I feel day-to-day, after more than 1 year. Being vegan may have helped me maintain a lower weight after I dropped several pounds last year, but it’s hard to know the counterfactual.
I didn’t know coming in that being vegan would be easy; I decided to try it out for 1 month and then stuck with it when I “learned” how to do it. There’s definitely a learning curve, but I’d say that for some people who get the hang of it, reason (1) in Michael’s comment genuinely applies.
I don’t think one year is enough time to observe effects. Anecdotically, I think (but am not sure) that I started to have problems after three years of being a vegetarian.
What were your problems?
Do you mean financial costs, or all net costs together, including potentially through time, motivation, energy, cognition? I think it’s reasonably likely that for many people, there are ~no real net (opportunity) costs, or that it’s actually net good (but if net good in those ways, then that would probably be a better reason than 1). Putting my thoughts in a footnote, because they’re long and might miss what you have in mind.[1]
Ya, that seems fair. If they had the option to just stop thinking and feeling bad about it and chose that over going veg, which is what my framing suggests they would do, then it seems the motivation is to feel better and get more time, not avoid harming animals through their diets. This would be like seeing someone in trouble, like a homeless person, and avoiding them to avoid thinking and feeling bad about them. This can be either selfish or instrumentally other-regarding, given opportunity costs.
If they thought (or felt!) the right response to the feelings is to just go veg and not to just stop thinking and feeling bad about it, then I would say they are in it to prevent harm, just guided by their feelings. And their feelings might not be very scope-sensitive, even if they make donation and career decisions in scope-sensitive ways. I think this is kind of what virtue ethics is about. Also potentially related: “Do the math, then burn the math and go with your gut”.
Financial/donations: It’s not clear to me that my diet is more expensive than if I were omnivorous. Some things I’ve substituted for animal products are cheaper and others are more expensive. I haven’t carefully worked through this, though. It’s also not clear that if it were more expensive, that I would donate more, because of how I decide how much to donate, which is based on my income and a vague sense of my costs of living, which probably won’t pick up differences due to diet (but maybe it does in expectation, and maybe it means donating less later, because I’ll have less money to donate later). If I budgeted more precisely, that would have time costs (which might not come out of work, anyway, though). And if I weren’t vegan, maybe I’d be less motivated to donate as much (although this is more like “self-signaling”, or altruism sharpens altruism).
Time: I doubt most veg*ns would work more hours (allowing more impactful work or more donations) if they weren’t managing or accommodating their veg*n diet. Time spent on the diet is small and probably doesn’t really come out of work hours. But this can depend on their particular circumstances. Maybe someone gets their blood tested more frequently because of their diet, and this specifically comes out of their work hours.
Time costs might also be incurred by others, like event or working space organizers (like you, as I think you’ve brought up before), and veg*ns may not appreciate this, or might (I’d guess correctly) think one more vegan makes ~no difference. Maybe it would be better if everyone agreed to eat whatever at events, and the animal products were offset by animal charity donations, or that time was just given back to organizers to work on other important things without any specific donation offset.
Effects on energy and cognition will vary by person. I think there are risks here people should make some effort to minimize (e.g. blood testing, supplements). That effort can come out of their time and donations, but that’s already accounted for above. There might be some remaining (expected) costs even after this. Or there could be net benefits, in case they end up with an overall healthier diet this way (and might not be motivated to do so except to go veg; it’s easier to avoid unhealthy foods you also object to morally).
I meant net costs all together, tough I agree that if you take into account motivation “net costs” becomes a kind of tricky concept, and many people can find it motivating, and that is important to think about, but also really doesn’t fit nicely into a harm-reducing framework.
I mean, being an onmivore would allow you to choose between more options. Generally having more options very rarely hurts you.
Overall I like your comment.
Ya, I guess the value towards harm reduction would be more indirect/instrumental in this case.
I think this is true of idealized rational agents with fixed preferences, but I’m much less sure about actual people, who are motivated in ways they wouldn’t endorse upon reflection and who aren’t acting optimally impartially even if they think it would be better on reflection if they did.
By going veg, you eliminate or more easily resist the motivation to eat more expensive animal products that could have net impartial opportunity costs. Maybe skipping (expensive) meats hurts you in the moment (because you want meat), but it saves you money to donate to things you think matter more. You’d be less likely to correctly — by reflection on your impartial preferences — skip the meat and save the money if you weren’t veg.
And some people are not even really open to (cheaper) plant-based options like beans and tofu, and that goes away going veg. That was the case for me. My attitude before going veg would have been irrational from an impartial perspective, just considering the $ costs that could be donated instead.
Of course, some people will endorse being inherently partial to themselves upon reflection, so eating animal products might seem fine to them even at greater cost. But the people inclined to cut out animal products by comparing their personal costs to the harms to animals probably wouldn’t end up endorsing their selfish motivation to eat animal products over the harms to animals.
The other side is that a veg*n is more motivated to eat the more expensive plant-based substitutes and go to vegan restaurants, which (in my experience) tend to be more expensive.
I’m not inclined to judge how things will shake out based on idealized models of agents. I really don’t know either way, and it will depend on the person. Cheap veg diets seem cheaper than cheap omni diets, but if people are eating enough plant-based meats, their food costs would probably increase.
Here are prices in Canadian $/kg of protein, for the versions of foods that seemed cheapest per kg of protein from Walmart Canada and Amazon Canada.
And then extra supplements for veg EAs.
For restaurants, as long as you avoid expensive plant-based substitutes and vegan-specific restaurants, I think veg options are usually cheaper. Of course, a veg EA will be tempted by these sometimes, too!
To be clear, I don’t think it’s that important to minimize the cost of your diet. Things like rent, vehicle, travel, and how often you eat out (when it doesn’t help you do more work) are probably more important if you want to save money.
I guess this is a bit pedantic, but you originally wrote “My best guess is that most people who are vegetarians, vegans or reducetarians, and are actually interested in scope-sensitivity, are explicitly doing so for signaling and social coordination reasons”. I think veg EAs are generally “actually interested in scope-sensitivity”, whether or not they’re thinking about their diets correctly and in scope-sensitive ~utilitarian terms. “The people who thought about it the most” might not be representative, and more representative motivations might be better described as “in it to prevent harm”, even if the motivations turn out to be not utilitarian, not appropriately scope-sensitive or misguided.
Happy to hear this from someone working on AI Safety.
Do you think EA/EA orgs should do something about it?
(Crypto has a similar problem. There are people genuinely interested in it but many people (most?) are just into it for the money)
I actually think it is a huge risk because when the shit hits the fan I want to know people are optimising for the right thing. It is like the old saying; “it is only a value if it costs you something”
Anyway, I’d like to think that orgs suss this out in the interview stage.
A psychological trap EAs might want to watch out for: don’t mistake indecision for wisdom.
Sometimes the better option is to just put something out into the world in a small format and get information back.
Of course the reverse works: don’t mistake decisiveness for bravery. But I don’t think this is a problem for EAs :)
People in EA are ridiculously helpful. I’ve been able to get a Careers Conference off the ground because people in other organisations are happy to help out in significant ways. In particular, Elliot Teperman and Bridget Loughhead from EA Australia, Greg Sadler from the Good Ancestors Project and Chris Leong. Several people have also volunteered to help at the Conference as well as on other tasks, such as running meetups and creating content.
In theory I’m running an organisation by myself (via an LTFF grant), but because it is nested within the EA/AIS space, I’ve got the support of a big network. Otherwise this whole project would be insane to start.
If someone is serious about doing Movement Building, but worries about doing it alone, this feels like important information.
A periodic reminder that you can just email politicians and then meet them (see screenshot below).
Another thing you can do is send comments proposed legislation on regulations.gov. I did so last week about a recent californian bill on open-sourcing model weights (now closed). In the checklist (screenshot below) they say: “the comment process is not a vote – one well supported comment is often more influential than a thousand form letters”. There are people much more qualified on AI risk than I over here, so in case you didn’t know, you might want to keep an eye on new regulation coming up. It doesn’t take much time and seems to have a fairly big impact.
I think it is good to have some ratio of upvoted/agreed : downvotes/disagreed posts in your portfolio. I think if all of your posts are upvoted/high agreeance then you’re either playing it too safe or you’ve eaten the culture without chewing first.
I think some kinds of content are uncontroversially good (e.g. posts that are largely informational rather than persuasive), so I think some people don’t have a trade-off here.
Good point. In that case the hypothetical user isn’t using it as a forum (i.e. for discourse)
as of comment, 6 agrees and 6 disagrees. perfect :)
[IMAGE] Extremely proud and excited to watch Greg Sadler (CEO of Good Ancestors Policy) and Soroush Pour (Co-Founder of Harmony Intelligence) speak to the Australian Federal Senate Select Committee on Artificial Intelligence. This is a big moment for the local movement. Also, check out who they ran after.
My previous take on writing to Politicians got numbers, so I figured I’d post the email I send below.
I am going to make some updates, but this is the latest version:
---
Hi [Politician]
My name is Yanni Kyriacos, I live in Coogee, just down the road from your electorate.
If you’re up for it, I’d like to meet to discuss the risks posed by AI. In addition to my day job building startups, I do community / movement building in the AI Safety / AI existential risk space. You can learn more about AI Safety ANZ by joining our Facebook group here or the PauseAI movement here. I am also a signatory of Australians for AI Safety—a group that has called for the Australian government to set up an AI Commission (or similar body).
Recently I worked with Australian AI experts (such as Good Ancestors Policy) in making a submission to the recent safe and response AI consultation process. In the letter, we called on the government to acknowledge the potential catastrophic and existential risks from artificial intelligence. More on that can be found here.
There are many immediate risks from already existing AI systems like ChatGPT or Midjourney, such as disinformation or improper implementation in various businesses. In the not-so-distant future, certain safety nets will need to be activated (such as a Universal Basic Income policy) in the event of mass unemployment due to displacement of jobs with robots and AI systems.
But of greatest concern is the speed at which we are marching towards AGI (artificial general intelligence) – systems that will have cognitive abilities at or above human level.
Half of AI researchers believe that there is a 10% or greater chance that the invention of artificial superintelligence will mean the end of humanity. Among AI safety scientists, this chance is estimated to be an average of 30%. And these levels of risk aren’t just a concern for people in the far-distant future, with prediction markets such as Metaculus showing these kinds of AI could be invented in the next term of government.
Notable examples of individuals sounding the alarm are Prof. Geoffrey Hinton and Prof. Yoshua Bengio, both Turing-award winners and pioneers of the deep learning methods that are currently achieving the most success. The existential risk of AI has been acknowledged by hundreds of scientists, the UN, the US and recently the EU.
To make a long story short: we don’t know how to align AI with the complex goals and values that humans have. When a superintelligent system is realised, there is a significant risk it will pursue a misaligned goal without us being able to stop it. And even if such a superhuman AI remains under human control, the person (or government) wielding such a power could use this to drastically, irreversibly change the world. Such an AI could be used to develop new technologies and weapons, manipulate masses of people or topple governments.
The advancements in the AI landscape have progressed much faster than anticipated. In 2020, it was estimated that an AI would pass university entrance exams by 2050. This goal was achieved in March 2023 by the system GPT-4 from OpenAI. These massive, unexpected leaps have prompted many experts to request a pause in AI development through an open letter to major AI companies. The letter has been signed over 33,000 times so far, including many AI researchers and tech figures.
Unfortunately, it seems that companies are not willing to jeopardise their competitive position by voluntarily halting development. A pause would need to be imposed by a government. Luckily, there seems to be broad support for slowing down AI development. A recent poll indicates that 63% of American support regulations to prevent AI companies from building superintelligent AI. At the national level, a pause is also challenging because countries have incentives to not fall behind in AI capabilities. That’s why we need an international solution.
The UK organised an AI Safety Summit on November 1st and 2nd at Bletchley Park. We hoped that during this summit, leaders would work towards sensible solutions that prevent the very worst of the risks that AI poses. As such I was excited to see that Australia signed the The Bletchley Declaration, agreeing that this risk is real and warrants coordinated international action. However, the recent policy statements by Minister Husic don’t seem to align with the urgency that experts are seeing. The last safe moment to act could be very soon.
The Summit has not yet produced an international agreement or policy. We have seen proposals being written by the US Senate, and even AI company CEOs have said there is “overwhelming consensus” that regulation is needed. But no proposal so far has seriously considered ways to slow down or prevent a superintelligent AI from being created. I am afraid that lobbying efforts by AI companies to keep regulation at a minimum are turning out to be highly effective.
It’s essential that the government follows through on its commitment at Bletchley Park to create a national or regional AI safety body. We have such bodies for everything from the risk of plane crashes to the risk of tsunamis. We urgently need one on ensuring the safety of AI systems.
Anyway, I’d love to discuss this more in person or via zoom if you’re in town soon.
Let me know what you think.
Cheers,,
Yanni
Thanks for sharing, Yanni, and it is really cool that you managed to get Australia’s Assistant Minister for Defence interested in creating an AI Safety Institute!
Did you mean to include a link?
The Metaculus’ question you link to involves meeting many conditions besides passing university exams:
[PHOTO] I sent 19 emails to politicians, had 4 meetings, and now I get emails like this. There is SO MUCH low hanging fruit in just doing this for 30 minutes a day (I would do it but my LTFF funding does not cover this). Someone should do this!
(Speaking as someone on LTFF, but not on behalf of LTFF)
How large of a constraint is this for you? I don’t have strong opinions on whether this work is better than what you’re funded to do, but usually I think it’s bad if LTFF funding causes people to do things that they think is less (positively) impactful!
We probably can’t fund people to do things that are lobbying or lobbying-adjacent, but I’m keen to figure out or otherwise brainstorm an arrangement that works for you.
Hey Linch, thanks for reaching out! Maybe send me your email or HMU here yannikyriacos@gmail.com
(I think you blanked out their name from their email footer but not from the bottom of the email, but not sure how anonymous you wanted to keep them.)
ffs
fixed. thanks mate :)
You should probably also blank their job title (which would make it easy to work out who they are) and their phone number (!)
Why am I so bad at this Stephen. Send help.
🫶
RIP to any posts on anything earnest over the last 48 hours. Maybe in future we don’t tag anything April Fools and it is otherwise a complete blackout on serious posts 😅
How about making the April Fool’s Day tag visible on the forum frontpage, like so?
Something(!) needs to be done. Otherwise, it’s just a mess for clarity and the communication of ideas.
I think the hilarity is in the confusion / click bait. Your idea would rob us of this! I think the best course of action is for anyone with a serious post to wait until April 3 :|
Not a solution to everything mentioned here- but a reminder that you can click “customize feed” at the top of the page and remove all posts tagged april fools.
nah let’s lean all the way in, for one day a year, the wild west out here.
Damn just had the idea of a “Who wants to be Fired?” post.
I think if you work in AI Safety (or want to) it is very important to be extremely skeptical of your motivations for working in the space. This applies to being skepticism of interventions within AI Safety as well.
For example, EAs (like most people!) are motivated to do things they’re (1) good at (2) see as high status (i.e. people very quietly ask themselves ‘would someone who I perceive as high status approve of my belief or action?‘). Based on this, I am worried that (1) many EAs find protesting AI labs (and advocating for a Pause in general) cringy and/or awkward (2) Ignore the potential impact of organisations such as PauseAI.
We might all literally die soon because of misaligned AI, so what I’m recommending is that anyone seriously considering AI Safety as a career path spends a lot of time on the question of ‘what is really motivating me here?’
fwiw i think this works in both directions—people who are “action” focussed probably have a bias towards advocacy / protesting and underweight the usefulness of research.
I have written 7 emails to 7 Politicians aiming to meet them to discuss AI Safety, and already have 2 meetings.
Normally, I’d put this kind of post on twitter, but I’m not on twitter, so it is here instead.
I just want people to know that if they’re worried about AI Safety, believe more government engagement is a good thing and can hold a decent conversation (i.e. you understand the issue and are a good verbal/written communicator), then this could be an underrated path to high impact.
Another thing that is great about it is you can choose how many emails to send and how many meetings to have. So it can be done on the side of a “day job”.
Big AIS news imo: “The initial members of the International Network of AI Safety Institutes are Australia, Canada, the European Union, France, Japan, Kenya, the Republic of Korea, Singapore, the United Kingdom, and the United States.”
https://www.commerce.gov/news/press-releases/2024/09/us-secretary-commerce-raimondo-and-us-secretary-state-blinken-announce
H/T @shakeel
Another shower thought—I’ve been vegetarian for 7 years and vegan for 6.5.
I find it very surprising and interesting how little it has nudged anyone in my friends / family circle to reduce their consumption of meat.
I would go as far as to say it has made literally no difference!
Keep in mind I’m surrounded by extremely compassionate and mostly left leaning people.
I also have relatively wide networks and I normally talk about animal welfare in a non annoying way.
You’d have thought in that time one person would have been nudged—nope!
My intention going vegan wasn’t as some social statement, but if you’d asked me 7 years later would there have been some social effect, I would have guessed yes.
Im surprised!
When we asked current vegetarians/vegans what influenced them to adopt their diet, personal conversations (primarily with friends/family) were among the top influences.
So your surprise/expectation seems reasonable! Of course, I don’t know whether it’s actually surprising, since presumably whether anyone actually converts depends on lots of other features of a given social network (do your networks contain a lot of people who were already vegetarian/vegan?).
Thanks for sharing that data David, I appreciate that :)
Do you think it can answer the question; If you’re the first in your family to switch to veganism/vegetarianism, how likely is it that another family member will follow suit within X years?
(Also, no, there aren’t any vegetarians or vegans among my friends or family).
FWIW I’d probably count myself as having ~ 30 close friends and family.
This also seems like good research for the GWWC pledge.
Thanks!
I’m afraid that. I think we’d need to know more about the respondents’ families and the order in which they adopted veganism/vegetarianism to assess that. I agree that it sounds like an interesting research question!
Maybe one day 😊
Have you tried cooking your best vegan recipes for others? In my experience sometimes people ask for the recipe and make it for themselves later, especially health-conscious people. For instance, I really like this vegan pumpkin pie that’s super easy to make: https://itdoesnttastelikechicken.com/easy-vegan-pumpkin-pie/
Thanks for the comment Sean. Because of me they’ve had to (seemingly gladly) eaten vegan-only meals dozens of times. My mum has even made me vegan dishes to take home on dozens of occasions. They almost all acknowledge the horror of factory farming. Veganism is just firmly bucketed a “yanni thing”.
That is interesting.
I think that most beings[1] who I’ve become very close with (I estimate this reference class is ≤10) have gone vegan in a way that feels like it was because of our discussions, though it’s possible they would have eventually otherwise, or that our discussions were just the last straw.
In retrospect, I wonder if I had first emanated a framing of taking morality seriously (such that “X is bad” → “I will {not do X} or try to stop X”). I think I also tended to become that close with beings who do already, and who are more intelligent/willing to reflect than what seems normal.
(I write ‘being’ because some are otherkin)
[GIF] A feature I’d love on the forum: while posts are read back to you, the part of the text that is being read is highlighted. This exists on Naturalreaders.com and would love to see it here (great for people who have wandering minds like me)
I agree with you, and so does our issue tracker. Sadly, it does seem a bit hard. Tagging @peterhartree as a person who might be able to tell me that it’s less hard than I think.
As someone who works with software engineers, I have respect for how simple-appearing things can actually be technically challenging.
For what it’s worth, I would find the first part of the issue (i.e. making the player “floating” or “sticky”) already quite useful, and it seems much easier to implement.
I think acting on the margins is still very underrated. For e.g. I think 5x the amount of advocacy for a Pause on capabilities development of frontier AI models would be great. I also think in 12 months time it would be fine for me to reevaluate this take and say something like ‘ok that’s enough Pause advocacy’.
Basically, you shouldn’t feel ‘locked in’ to any view. And if you’re starting to feel like you’re part of a tribe, then that could be a bad sign you’ve been psychographically locked in.
I’m pretty confident that Marketing is in the top 1-3 skill bases for aspiring Community / Movement Builders.
When I say Marketing, I mean it in the broad sense it used to mean. In recent years “Marketing” = “Advertising”, but I use the classic Four P’s of Marketing to describe it.
The best places to get such a skill base is at FMCG / mass marketing organisations such as the below. Second best would be consulting firms (McKinsey & Company):
Procter & Gamble (P&G)
Unilever
Coca-Cola
Amazon
1. Product
- What you’re selling (goods or services)
- Features and benefits
- Quality, design, packaging
- Brand name and reputation
- Customer service and support
2. Price
- Retail/wholesale pricing
- Discounts and promotions
- Payment terms
- Pricing strategy (premium, economy, etc.)
- Price comparison with competitors
3. Place (Distribution)
- Sales channels
- Physical/online locations
- Market coverage
- Inventory management
- Transportation and logistics
- Accessibility to customers
4. Promotion
- Advertising
- Public relations
- Sales promotions
- Direct marketing
- Digital marketing
- Personal selling
The Australian Federal Government is setting up an “AI Advisory Body” inside the Department of Industry, Science and Resources.
IMO that means Australia has an AISI in:
9 months (~ 30% confidence)
12 months (~ 50%)
18 months (~ 80%)
That link doesn’t work for me fyi
Interestingly, they’ve taken the page down and the open positions on LinkedIn.
My guess is they announced too early.
Nice work—obviously it’s hard to attribute but I’m sure the combined lobbying of all the EAs working on this in Aus have contributed to this one.
I’d assign a 75% probably that we’ve made it happen somewhat* sooner than it otherwise would have.
*Maybe 6 months? I honestly have no clue.
Yesterday Greg Sadler and I met with the President of the Australian Association of Voice Actors. Like us, they’ve been lobbying for more and better AI regulation from government. I was surprised how much overlap we had in concerns and potential solutions:
1. Transparency and explainability of AI model data use (concern)
2. Importance of interpretability (solution)
3. Mis/dis information from deepfakes (concern)
4. Lack of liability for the creators of AI if any harms eventuate (concern + solution)
5. Unemployment without safety nets for Australians (concern)
6. Rate of capabilities development (concern)
They may even support the creation of an AI Safety Institute in Australia. Don’t underestimate who could be allies moving forward!
Flaming hot take: I wonder if some EAs suffer from Scope Oversensitivity—essentially the inverse of the identifiable victim effect. Take the animal welfare vs global health debate: are we sometimes biased by the sheer magnitude of animal suffering numbers, rather than other relevant factors? Just as the identifiable victim effect leads people to overweight individual stories, maybe we’re overweighting astronomical numbers.
EAs pride themselves on scope sensitivity to combat emotional biases, but taken to an extreme, could this create its own bias? Are we sometimes too seduced by bigger numbers = bigger problem? The meta-principle might be that any framework, even one designed to correct cognitive biases, needs wisdom and balance to avoid becoming its own kind of distortion.
Scope insensitivity has some empirical backing—e.g. the helping birds study—and some theorised mechanisms of action, e.g. people lacking intuitive understanding of large numbers.
Scope oversensitivity seems possible in theory, but I can’t think of any similar empirical or theoretical reasons to think it’s actually happening.
To the extent that you disagree, it’s not clear to me whether it’s because you and I disagree on how EAs weight things like animal suffering, or whether we disagree on how it ought to be weighted. Are you intending to cast doubt on the idea that a problem that is 100x as large is (all else equal) 100x more important, or are you intending to suggest that EAs treat it as more than 100x as important?
Upvoted and disagree-voted.
Do you have any particular reason to believe that EAs overweight large problems?
EAs donate much more to global health than to animal welfare. Do you think the ratio should be even higher still?
The general public wants frontier AI models regulated and there doesn’t seem to be grassroots focussed orgs attempting to capture and funnel this energy into influencing politicians. E.g. via this kind of activity. This seems like massively low hanging fruit. An example of an organisation that does this (but for GH&W) is Results Australia. Someone should set up such an org.
My impression is that PauseAI focusses more on media engagement + protests, which I consider a good but separate thing. Results Australia, as an example, focuses (almost exclusively) on having concerned citizens interacting directly with politicians. Maybe it would be a good thing for orgs to focus on separate things (e.g. for reasons of perception + specialisation). I lean in this direction but remain pretty unsure.
Founder of PauseAI here. I know our protests are the most visible, but they are actually a small portion of what we do. People always talk about the protests, but I think we actually had most results through invisible volunteer lobbying. Personally, I’ve spent way more time sending emails to politicians and journalists, meeting them and informing them of the issues. I wrote an Email Builder to get volunteers to write to their representatives, gave multiple workshops on doing so, and have seen many people (including you of course!) take action and ask for feedback in the discord.
I think combining both protesting and volunteer lobbying in one org is very powerful. It’s not a new idea of course—orgs like GreenPeace have been using this strategy for decades. The protests create visibility and attention, and the lobbying gives the important people the right background information. The protests encourage more volunteers to join and help out, so we get more volunteer lobbyists! In my experience the protests also help with getting an in with politicians—it creates a recognizable brand that shows you represent a larger group.
Is anyone in the AI Governance-Comms space working on what public outreach should look like if lots of jobs start getting automated in < 3 years?
I point to Travel Agents a lot not to pick on them, but because they’re salient and there are lots of them. I think there is a reasonable chance in 3 years that industry loses 50% of its workers (3 million globally).
People are going to start freaking out about this. Which means we’re in “December 2019” all over again, and we all remember how bad Government Comms were during COVID.
Now is the time to start working on the messaging!
I believe we should worry about the pathology of people like Sam Altman, mostly because of what they’re trying to achieve, without the consent of all governments and their respective populace. I found a tweet that summarises my own view better than I can. The pursuit of AGI isn’t normal or acceptable and we should call it out as such.
Do you have an example of people disagreeing with you? When I made similar points before, I think they’ve been received relatively positively.
I’ll dig it up. Could very well be my delivery is off putting.
You know what, I’m gonna rewrite this quick take so it is about the tweet. Currently it is unnecessarily whiny / self-interested.
Remember: EA institutions actively push talented people into the companies making the world changing tech the public have said THEY DONT WANT. This is where the next big EA PR crisis will come from (50%). Except this time it won’t just be the tech bubble.
Is this about the safety teams at capabilities labs?
If so, I consider it a non-obvious issue, whether pushing a talented people into an AI safety role at, e.g., DeepMind is a bad thing. If you think that is a bad thing, consider providing a more detailed argument, and writing a top-level post explaining your view.
If, instead, this is about EA institutions pushing people into capabilities roles, consider naming these concrete examples. As an example, 80k has a job advertising a role as a prompt engineer at Scale AI. That does not seem to be a very safety-focused role, and it is not clear how 80k wants to help prevent human extinction with that job ad.
Thanks for the comment :) I think expecting this detail is normal from someone on the EA Forum. If you think that’s the way public relations are going to work, which is my main point, then we’re disagreeing.
So I did a quick check today—I’ve sent 19 emails to politicians about AI Safety / x-risk and received 4 meetings. They’ve all had a really good vibe, and I’ve managed to get each of them to commit to something small (e.g. email XYZ person about setting up an AI Safety Institute). I’m pretty happy with the hit rate (4/19). I might do another forum quick take once I’ve sent 50.
I expect (~ 75%) that the decision to “funnel” EAs into jobs at AI labs will become a contentious community issue in the next year. I think that over time more people will think it is a bad idea. This may have PR and funding consequences too.
My understanding is that this has been a contentious issue for many years already.
80,000 hours wrote a response to this last year, Scott Alexander had written about it in 2022, and Anthropic split from OpenAI in 2021. Do you mean that you expect this to become significantly more salient in the next year?
Thanks for the reply Lorenzo! IMO it is going to look VERY weird seeing people continue to leave labs while EA fills the leaky bucket.
Yeah my prediction lacked specificity. I expect it to become quite heated. I’m imagining 10+ posts on the topic next year with a total of 100+ comments. That’s just on the forum.
I’d probably bet against this happening FWIW. Maybe a Manifold market?
Also, 100+ comments on the forum might not mean it’s necessarily “heated”—a back and forth between two commenters can quickly increase the tally on any subject so that part might also need specifying further.
I spent about 30 seconds thinking about how to quantify my prediction. I’m trying to point at something vague in a concrete way but failing. This also means that I don’t think it is worth my time making it more concrete. The initial post was more of the “I am pretty confident this will be a community issue, just a heads up”.
Seems reasonable :)
Be the meme you want to see in the world (screenshot).
What is your best guess of the overall impact of 80k interviewing AI labs on their podcast + listing AI lab vacant roles?
Poll: https://www.guidedtrack.com/programs/qj0ykwn/run
@Yonatan Cale
My long thoughts:
1. 80k don’t claim to only advertise impactful jobs
They also advertise jobs that help build career impact, and they’re not against posting jobs that cause harm (and it’s often/always not clear which is which). See more in this post.
They sometimes add features like marking “recommended orgs” (which I endorse!), and sometimes remove those features ( 😿 ).
2. 80k’s career guide about working at AI labs doesn’t dive into “which lab”
See here. Relevant text:
I think [link to comment] the “which lab” question is really important, and I’d encourage 80k to either be opinionated about it, or at least help people make up their mind somehow, not just leave people hanging on “which lab” while also often recommending people go work at AI labs, and also mentioning that often that work is net-negative and recommending reducing the risk by not working “in certain positions unless you feel awesome about the lab”.
[I have longer thoughts on how they could do this, but my main point is that it’s (imo) an important hole in their recommendation that might be hidden from many readers]
3. Counterfactual / With great power comes great responsibility
If 80k wouldn’t do all this, should we assume there would be no job board and no guides?
I claim that something like a job board has critical mass: Candidates know the best orgs are there, and orgs know the best candidates are there.
Once there’s a job board with critical mass, it’s not trivial to “compete” with it.
But EAs love opening job boards. A few new EA job boards pop up every year. I do think there would be an alternative. And so the question seems to me to be—how well are 80k using their critical mass?
4. What results did 80k’s work actually cause?
First of all: I don’t actually know, and if someone from 80k would respond, that would be way better than my guess.
Still, here’s my guess, which I think would be better than just responding to the poll:
Lots of engineers who care about AI Safety but don’t have a deep understanding of it (and not much interest in spending months to learn) - go work at AI labs.
Is this positive, because now people “care” in the labs, or negative because the labs have a much easier time hiring people who basically go and do their job? This is seriously a hard question, but I’d guess “negative” (and I think 80k would agree but I’m not sure)
I wouldn’t be surprised if 80k are directly responsible for a few very important hires.
For example, I think the CISO (head of security) of Anthropic used to run the security of Chrome. I’m VERY happy Anthropic hired such a person, I think infosec is really important for AI labs, and I wouldn’t be surprised if 80k had something to do with this, and if not this—then maybe some other similar role.
I think this is very positive, and maybe more important than all the rest.
[I need to go, my comment seems incomplete but I hope it’s still somewhat helpful so posting. I’m still not sure how to vote!]
And there’s always the other option that I (unpopularly) believe in—that better publicly available AI capabilities are necessary for meaningful safety research, thus AI labs have contributed positively to the field.
NotebookLM is basically magic. Just take whatever Forum post you can’t be bothered reading but know you should and use NotebookLM to convert it into a podcast.
It seems reasonable that in 6 − 12 months there will be a button inside each Forum post that converts said post into a podcast (i.e. you won’t need to visit NotebookLM to do it).
It breaks my heart when I see eulogy posts on the forum. And while I greatly appreciate people going to the effort of writing them (while presumably experiencing grief), it still doesn’t feel like enough. We’re talking about people that dedicated their lives to doing good, and all they get is a post. I don’t have a suggestion to address this ‘problem’, and some may even feel that a post is enough, but I don’t. Maybe there is no good answer and death just sucks. I dunno.
I think https://www.wakingup.com/ should be considered for effective organisation status. It donates 10% of revenue to the most effective causes and I think reaching nondual states of awakening could be one of the most effective ways for people in rich countries to improve their wellbeing.
Related: https://www.clearerthinking.org/post/can-you-experience-enlightenment-through-sam-harris-waking-up-meditation-app
Also related (though more tangentially): https://podcast.clearerthinking.org/episode/167/michael-taft-and-jeremy-stevenson-glimpses-of-enlightenment-through-nondual-meditation/
A piece of career advice I’ve given a few times recently to people in AI Safety, which I thought worth repeating here, is that AI Safety is so nascent a field that the following strategy could be worth pursuing:
1. Write your own job description (whatever it is that you’re good at / brings joy into your life).
2. Find organisations that you think need thing that job but don’t yet have it. This role should solve a problem they either don’t know they have or haven’t figured out how to solve.
3. Find the key decision maker and email them. Explain the (their) problem as you see it and how having this role could fix their problem. Explain why you’re the person to do it.
I think this might work better for mid-career people, but, if you’re just looking to skill up and don’t mind volunteering, you could adapt this approach no matter what stage you’re at in your career.
I often see EA orgs looking to hire people to do Z, where applicants don’t necessarily need experience doing Z to be successful.
E.g. instead of saying “must have minimum 4 years as an Operations Manager”, they say “we are hiring for an Operations Manager. But you don’t need experience as an Operations Manager, as long as you have competencies/skills (insert list) A—F.”
This reminds of when Bloomberg spent over $10M training a GPT-3.5 class AI on their own financial data, only to find that GPT-4 beat it on almost all finance tasks.
They thought proprietary data would perform better, but it turns out the largest models won (at the time at least).
I worry that saying “4 years of experience as Operations Manager” is less important than the “competencies A—F” we’ve identified, makes the same mistake.
You’re gambling on having found the secret sauce. But as usual the more targeted the intervention, the more likely it is to miss.
If someone claims to have competencies A-F required to be an operation manager, they should have some way of proving that they actually have competencies A-F. A great way for them to prove this is to have 4 years of experience as Operations Manager. In fact, I struggle to think of a better way to prove this, and obviously a person with experience is preferable to someone with no experience for all kind of reasons (they have encountered all the little errors, etc). As a bonus, somebody outside of your personal bubble has trusted this person.
EA should be relying on more subject matter experience, not less.
Something bouncing around my head recently … I think I agree with the notion that “you can’t solve a problem at the level it was created”.
A key point here is the difference between “solving” a problem and “minimising its harm”.
Solving a problem = engaging with a problem by going up a level from which is was createwd
Minimising its harm = trying to solve it at the level it was created
Why is this important? Because I think EA and AI Safety have historically focussed (and has their respective strengths in) harm-minimisation.
This applies obviously the micro. Here are some bad examples:
Problem: I’m experiencing intrusive + negative thoughts
Minimising its harm: engage with the thought using CBT
Attempting to solve it by going meta: apply meta cognitive therapy, see thoughts as empty of intrinsic value, as farts in the wind
Problem: I’m having fights with my partner about doing the dishes
Minimising its harm: create a spreadsheet and write down every thing each of us does around the house and calculate time spent
Attempting to solve it by going meta: discuss our communication styles and emotional states when frustration arises
But I also think this applies at the macro:
Problem: People love eating meat
Minimising harm by acting at the level the problem was created: asking them not to eat meat
Attempting to solve by going meta: replacing the meat with lab grown meat
Problem: Unaligned AI might kill us
Minimising harm by acting at the level the problem was created: understand the AI through mechanistic interpretability
Attempting to solve by going meta: probably just Governance
To me, your examples at the micro level don’t make the case that you can’t solve a problem at the level it’s created. I’m agnostic as to whether CBT or meta-cognitive therapy is better for intrusive thoughts, but lots of people like CBT; and as for ‘doing the dishes’, in my household we did solve the problem of conflicts around chores by making a spreadsheet. And to the extent that working on communication styles is helpful, that’s because people (I’d claim) have a problem at the level of communication styles.
This is an extremely “EA” request from me but I feel like we need a word for people (i.e. me) who are Vegans but will eat animal products if they’re about to be thrown out. OpportuVegan? UtilaVegan?
Freegan
I think the term I’ve heard (from non-EAs) is ‘freegan’ (they’ll eat it if it didn’t cause more animal products to be purchased!)
This seems close enough that I might co-opt it :)
https://en.wikipedia.org/wiki/Freeganism
If you predictably do this, you raise the odds that people around you will cook some/ buy some extra food so that it will be “thrown out”, or offer you food they haven’t quite finished (and that they’ll replace with a snack later.
So I’d recommend going with “Vegan” as your label, for practical as well as signalling reasons.
Yeah this is a good point, which I’ve considered, which is why I basically only do it at home.
I am 90% sure that most AI Safety talent aren’t thinking hard enough about what Neglectedness. The industry is so nascent that you could look at 10 analogous industries, see what processes or institutions are valuable and missing and build an organisation around the highest impact one.
The highest impact job ≠ the highest impact opportunity for you!
I feel like you’re making this sound simple when I’d expect questions like this to involve quite a bit of work, and potentially skills and expertise that I wouldn’t expect people at the start of their careers to have yet.
Do you have any specific ideas for something that seems obviously missing to you?
AI Safety (in the broadest possible sense, i.e. including ethics & bias) is going be taken very seriously soon by Government decision makers in many countries. But without high quality talent staying in their home countries (i.e. not moving to UK or US), there is a reasonable chance that x/c-risk won’t be considered problems worth trying to solve. X/c-risk sympathisers need seats at the table. IMO local AIS movement builders should be thinking hard about how to either keep talent local (if they’re experiencing brain drain) OR increase the amount of local talent coming into x/c-risk Safety, such that outgoing talent leakage isn’t a problem.
What types of influence do you think governments from small, low influence countries will be able to have?
For example, the NZ government—aren’t they price-takers when it comes to AI regulation? If you’re not a significant player, don’t have significant resources to commit to the problem, and don’t have any national GenAI companies—how will they influence the development trajectory of AI?
That’s one way, but countries’s (and other body’s) concerns are made up of both their citizens/participants concerns and other people’s concerns. one is valued differently from the other, but there’s other ways of keeping relevant actors away from not thinking its a problem. (namely, utilizing the UN.) you make a good point though.
I beta tested a new movement building format last night: online networking. It seems to have legs.
V quick theory of change:
> problem it solves: not enough people in AIS across Australia (especially) and New Zealand are meeting each other (this is bad for the movement and people’s impact).
> we need to brute force serendipity to create collabs.
> this initiative has v low cost
quantitative results:
> I purposefully didn’t market it hard because it was a beta. I literally got more people that I hoped for
> 22 RSVPs and 18 attendees
> this says to me I could easily get 40+
> average score for below question was 7.27, which is very good for a beta test
I used Zoom, which was extremely clunky. These results suggest to me I should;
> invest in software designed for this use case, not zoom
> segment by career stream (governance vs technical) and/or experience (beginner vs advanced)
> run it every second month
I have heaps of qualitative feedback from participants but don’t have time to share it here.
Email me if interested: yanni@aisafetyanz.com.au
Two jobs in AI Safety Advocacy that AFAICT don’t exist, but should and probably will very soon. Will EAs be the first to create them though? There is a strong first mover advantage waiting for someone −
1. Volunteer Coordinator—there will soon be a groundswell from the general population wanting to have a positive impact in AI. Most won’t know how to. A volunteer manager will help capture and direct their efforts positively, for example, by having them write emails to politicians
2. Partnerships Manager—the President of the Voice Actors guild reached out to me recently. We had a very surprising number of cross over in concerns and potential solutions. Voice Actors are the canary in the coal mine. More unions (etc) will follow very shortly. I imagine within 1 year there will be a formalised group of these different orgs advocating together.
I think the average EA worries too much about negative PR related to EA. I think this is a shame because EA didn’t get to where it is now because it concerned itself with PR. It got here through good-old-fashioned hard work (and good thinking ofc).
Two examples:
1. FTX.
2. OpenAI board drama.
On the whole, I think there was way too much time spent thinking and talking about these massive outliers and it would have been better if 95%+ of EAs put their head down and did what they do best—get back to work.
I think it is good to discuss and take action on the issues at a concrete level, but not to talk about how this affects “what people think of EA”. EA overindexing on conscientiousness can have negative consequences!
What is your “Pens Down” moment?
“Pens Down” to mean ‘Artificial Super Intelligence in my opinion is close enough that it no longer makes sense to work on whatever else I’m currently working on, because we’re about to undergo radical amounts of change very soon/quickly’.
For me, it is probably when we have something as powerful as GPT-4 except it is agentic and costs less than $100 / month. So, that looks like a digital personal assistant that can execute an instruction like “have a TV delivered for me by X date, under Y price and organise installation and wall mounting.”
This is obviously a question mainly for people who don’t work full time on AI Safety.
I don’t know if this can be answered in full-generality.
I suppose it comes down to things like:
• Financial runway/back-up plans in case your prediction is wrong
• Importance of what you’re doing now
• Potential for impact in AI safety
I agree. I think it could be a useful exercise though to make the whole thing (ASI) less abstract.
I find it hard to reconcile that (1) I think we’re going to have AGI soon and (2) I haven’t made more significant life changes.
I don’t buy the argument that much shouldn’t change (at least, in my life).
Happy to talk that through if you’d like, though I’m kind of biased, so probably better to speak to someone who doesn’t have a horse in the race.
I suppose it is plausible for a person to never have a “Pens Down” moment if;
There is a FOOM
They feel they won’t be able to positively contribute to making ASI safe / slowing it down
I’m somewhat worried we’re only 2-3 years from this, FWIW. I’d give it a ~ 25% chance.
A potential failure mode of 80k recommending EAs work at AI labs:
80k promotes a safety related job within a leading AI lab.
80k’s audience (purposefully) skews to high prospect candidates (HPC) - smarter, richer, better connected vs average.
HPC applies for and gets safety role within AI lab.
HPC candidate stays at the lab but moves roles.
Now we have a smart, rich, well connected person no longer in safety but in capabilities.
I think this is sufficiently important / likely that 80k should consider tracking these people over time to see if this is a real issue.
Thanks Yanni, I think a lot of people have been concerned about this kind of thing.
I would be surprised if 80,000 hours isn’t already tracking this or something like it—perhaps try reaching out to them directly, you might get a better response that way
Three million people are employed by the travel (agent) industry worldwide. I am struggling to see how we don’t lose 80%+ of those jobs to AI Agents in 3 years (this is ofc just one example). This is going to be an extremely painful process for a lot of people.
I think the greater threat to travel agents was the rise of self-planning travel internet sites like skyscanner and booking.com which make them mostly unnecessary. If travel agents have survived that, I don’t see how they wouldn’t survive LLM’s: presumably one of the reasons they survived is the preference for talking in person to a human being.
Yep, I know people who even opt against self-checkout because they enjoy things like a sense of community! I don’t claim AI isn’t going to greatly disrupt a number of jobs, but service jobs in particular seem like some of the most resistant, especially if people get wealthier from AI and have more money to spend on niceties.
Avatars are going to be extremely high fidelity. It will feel like a zoom call with a colleague, except you’re booking travel.
I’m not sure it going to feel like a zoom call with a colleague any time soon. That’s a pretty high bar IMO that we ain’t anywhere near yet. Many steps aren’t yet there which include
(I’m probably misrepresenting a couple of these due to lack of expertise but something like...)
1) Video quality (especially being rendered real time)
2) Almost insta-replying
3) Facial warmth and expressions
4) LLM sounding exactly like a real person (this one might be closest)
5) Reduction in processing power required for this to be a norm. Having thousands of these going simultaneously is going to need a lot. (again less important)
I would bet against Avatars being this high fidelity in 3 years a in common use because I think LLM progress is tailing off and there are multiple problems to be solved to get there—but maybe I’m a troglodyte...
That would not contribute to my friends’ sense of community or trust in other people (or, for that matter, the confidence that if a real service person gives you a bad experience or advice, that you have a path to escalate it)
I always get downvoted when I suggest that (1) if you have short timelines and (2) you decide to work at a Lab then (3) people might hate you soon and your employment prospects could be damaged.
What is something obvious I’m missing here?
One thing I won’t find convincing is someone pointing at the Finance industry post GFC as a comparison.
I believe the scale of unemployment could be much higher. E.g. 5% ->15% unemployment in 3 years.
If you have short timelines, I feel like your employment prospects in a fairly specific hypothetical should be fairly low on your priority list compared to ensuring AI goes well? Further, if you have short timelines, knowing about AI seems likely to be highly employable and lab experience will be really useful (until AI can do it better than us...)
I can buy “you’ll be really unpopular by affiliation and this will be bad for your personal life” as a more plausible argument
Hi! Some thoughts:
“If you have short timelines, I feel like your employment prospects in a fairly specific hypothetical should be fairly low on your priority list compared to ensuring AI goes well?”
- Possibly! I’m mostly suggesting that people should be aware of this possibility. What they do with that info is obvs up to them
”if you have short timelines, knowing about AI seems likely to be highly employable and lab experience will be really useful”
- The people I’m pointing to here (e.g. you) will be employable and useful. Maybe we’re talking passed each other a bit?
Why do you believe this? Being an expert in AI is currently very employable, and seems all the more so if AI is automating substantial parts of the economy
Ugh I wrote the wrong thing here, my bad (will update comment).
Should have said
“—The people I’m pointing to here (e.g. you) will be employable and useful. Maybe we’re talking passed each other a bit?”
I’m very confused. Your top level post says “your employment prospects could be damaged.” I am explaining why I think that the skillset of lab employees will be way more employable. Is your argument that the PR will outweigh that, even though it’s a major factor?
Financial workers’ careers are not directly affected very much by populist prejudice because 1) they provide a useful service and ultimately people pay for services and 2) they are employed by financial firms who do not share that prejudice. Likewise in short timeline worlds AI companies are probably providing a lot of services to the world, allowing them to pay high compensation, and AI engineers would have little reason to want to work elsewhere. So even if they are broadly disliked (which I doubt) they could still have very lucrative careers.
Obviously its a different story if antipathy towards the financial sector gets large enough to lead to a communist takeover.
During the financial crisis U-3 rose from 5% to 10% over about 2 years, suggesting on this metric at least you should think it is at least comparable (2.5%/year vs 3.3%/year).
Just on that last prediction—that’s an extraordinary claim given that the U.S. unemployment rate has only increased 0.7% since ChatGPT was released nearly 2 years ago (harder to compare again GPT-3, which was released in 2020, but you could say the same about the economic impacts of recent LLMs more broadly). If you believe we’re on the precipice of some extraordinary increase, why stop at 15%? Or if we’re not on that precipice, what’s the mechanism for that 10% jump?
Heya mate, it is an extraordinary claim, depending on how powerful you expect the models to be and in what timeframe!
“U.S. unemployment rate has only increased 0.7% since ChatGPT was released nearly 2 years ago”
“what’s the mechanism for that 10% jump?”
I’m pretty confident that a majority of the population will soon have very negative attitudes towards big AI labs. I’m extremely unsure about what impact this will have on the AI Safety and EA communities (because we work with those labs in all sorts of ways). I think this could increase the likelihood of “Ethics” advocates becoming much more popular, but I don’t know if this necessarily increases catastrophic or existential risks.
Anecdotally I feel like people generally have very low opinion of and understanding of the finance sector, and the finance sector mostly doesn’t seem to mind. (There are times when there’s a noisy collision, e.g. Melvin Capital vs. GME, but they’re probably overall pretty rare.)
It’s possible / likely that AI and tech companies targeting mass audiences with their product are more dependent on positive public perception, but I suspect that the effect of being broadly hated is less strong than you’d think.
I think the amount of AI caused turmoil I’m expecting in 3-5 years might be much higher than you, which accounts for most of this difference.
The public already has a negative attitude towards the tech sector before the AI buzz. in 2021 45% of americans had a somewhat or very negative view of tech companies.
I doubt the prevalence of AI is making people more positive towards the sector given all the negative publicity over plagarism, job loss, and so on. So I would guess the public already dislikes AI companies (even if they use their products), and this will probably increase.
Yeah the problem with some surveys is they measure prompted attitudes rather than salient ones.
Can you elaborate on what makes you so certain about this? Do you think that the reputation will be more like that of Facebook or that of Big Tobacco? Or will it be totally different?
Basically, I think there is a good chance we have 15% unemployment rates in less than two years caused primarily by digital agents.
Totally different. I had a call with a voice actor who has colleagues hearing their voices online without remuneration. Tip of the iceberg stuff.
More thoughts on “80,000 hours should remove OpenAI from the Job Board”
- - - -
A broadly robust heuristic I walk around with in my head is “the more targeted the strategy, the more likely it will miss the target”.
Its origin is from when I worked in advertising agencies, where TV ads for car brands all looked the same because they needed to reach millions of people, while email campaigns were highly personalised.
The thing is, if you get hit with a tv ad for a car, worst case scenario it can’t have a strongly negative effect because of its generic content (dude looks cool while driving his 4WD through rocky terrain). But when an email that attempts to be personalised misses the mark, you feel it (Hi {First Name : Last Name}!).
Anyway, effective altruism loves targeting. It is virtually built from the ground up on it (RCTs and all that, aversion to PR etc).
Anyway, I am thinking about this in relation to the recent “80,000 hours should remove OpenAI from the Job Board” post.
80k is applying an extremely targeted strategy by recommending people work in some jobs at the workplaces creating the possible death machines (caveat 1, caveat 2, caveat 3, caveatn).
There are so many failure modes involved with those caveats that makes it virtually impossible to not go catastrophically poorly in at least a few instances and take the mental health of a bunch of young and well-meaning EAs with it.
A judgement I’m attached to is that a person is either extremely confused or callous if they work in capabilities at a big lab. Is there some nuance I’m missing here?
Some of them have low p(doom from AI) and aren’t longtermists, which justifies the high-level decision of working on an existentially dangerous technology with sufficiently large benefits.
I do think their actual specific actions are not commensurate with the level of importance or moral seriousness that they claim to attach to their work, given those stated motivations.
Thanks for the comment Linch! Just to spell this out:
“Some of them have low p(doom from AI) and aren’t longtermists, which justifies the high-level decision of working on an existentially dangerous technology with sufficiently large benefits. ”
I would consider this acceptable if their p(doom) was ~ < 0.01%
I find this pretty incredulous TBH
“I do think their actual specific actions are not commensurate with the level of importance or moral seriousness that they claim to attach to their work, given those stated motivations.”
I’m a bit confused by this part. Are you saying you believe the importance / seriousness the person claims their work has is not reflected in the actions they actually take? In what way are you saying they do this?
I dunno man, at least some people think that AGI is the best path to curing cancer. That seems like a big deal! If you aren’t a longtermist at all, speeding up the cure for cancer by a year is probably worth quite a bit of x-risk.
Yes.
Lab people shitpost, lab people take competition extremely seriously (if your actual objective is “curing cancer,” it seems a bit discordant to be that worried that Silicon Valley Startup #2 will cure cancer before you), people take infosec not at all seriously, the bizarre cultish behavior after the Sam Altman firing, and so forth.
In my experience, they’re mostly just impulsive and have already written their bottom line (“I want to work on AI projects that look cool to me”), and after that they come up with excuses to justify this to themselves.
Hi! Thanks for your comment :)
I basically file this under “confused”.
Interesting, in your ontology I’d much more straightforwardly characterize it as “callous”
Sorry, I will explain further for clarity. I am Buddhist and in that philosophy the phrase “confused” has an esoteric usage. E.g. “that person suffers because they’re confused about the nature of reality”. I think this lab person is confused about how much suffering they’re going to create for themselves and others (i.e. it hasn’t been seen or internalised).
Yes. I think most people working on capabilities at leading labs are confused or callous (or something similar, like greedy or delusional), but definitely not all. And personally, I very much hope there are many safety-concerned people working on capabilities at big labs, and am concerned about the most safety-concerned people feeling the most pressure to leave, leading to evaporative cooling.
Reasons to work on capabilities at a large lab:
To build career capital of the kind that will allow you to have a positive impact later. E.g. to be offered relevant positions in government
To positively influence the culture of capabilities teams or leadership at labs.
To be willing and able to whistleblow bad situations (e.g. seeing emerging dangerous capabilities in new models, the non-disparagement stuff).
[maybe] to earn to give (especially if you don’t think you’re contributing to core capabilities)
To be clear, I expect achieving the above to be infeasible for most people, and it’s important for people to not delude themselves into thinking they’re having a positive impact to keep enjoying a lucrative, exciting job. But I definitely think there are people for whom the above is feasible and extremely important.
Another way to phrase the question is “is it good for all safety-concerned people to shun capabilities teams, given (as seems to be the case) that those teams will continue to exist and make progress by default?” And for me the strong answer is “yes”. Which is totally consistent with wanting labs to pause and thinking that just contributing to capabilities (on frontier models) in expectation is extremely destructive.
Thanks so much for your thoughtful comment! I appreciate someone engaging with me on this rather than just disagree ticking. Some thoughts:
1. Build Career Capital for Later Impact:
I think this depends somewhat on what your AGI timelines are. If they’re short, you’ve wasted your time and possibly had a net-negative effect.
I think there is a massive risk of people entering an AGI lab for career capital building reasons, and then post rationalising their decision to stay. Action changes attitudes faster than attitude changes actions after all.
2. Influence the Culture:
Multiple board members attempted this at OpenAI got fired, what chance does a single lower level employee?
I have tried changing corporate culture at multiple organisations and it is somewhere between extremely hard and impossible (more on that below).
3. Be Prepared to Whistle-Blow:
I was a Whistle-Blower during my time at News Corp. It was extremely difficult. I simply do not expect a meaningful number of people to be able to do this. You have to be willing to turn your life upside down.
This can be terrible for a person’s mental health. We shouldn’t be vary careful openly recommending this as a reason to stay at a Lab.
As mentioned above, I expect a greater number of people to turn into converts than would-be Whistle-Blowers
I think if you’re the kind of person who goes into a Lab for Career Capital that makes you less likely to be a Whistle-Blower TBH.
4. Be Prepared to Whistle-Blow:
Sorry, you can make good money elsewhere.
(1) I agree if your timelines are super short, like <2yrs, it’s probably not worth it. I have a bunch of probability mass on longer timelines, though some on really short ones
Re (2), my sense is some employees already have had some of this effect (and many don’t. But some do). I think board members are terrible candidates for changing org culture; they have unrelated full-time jobs, they don’t work from the office, they have different backgrounds, most people don’t have cause to interact with them much. People who are full-time, work together with people all day every day, know the context, etc., seem more likely to be effective at this (and indeed, I think they have been, to some extent in some cases)
Re (3), seems like a bunch of OAI people have blown the whistle on bad behavior already, so the track record is pretty great, and I think them doing that has been super valuable. And 1 whistleblower seems much better than several converts is bad. I agree it can be terrible for mental health for some people, and people should take care of themselves.
Re (4), um, this is the EA Forum, we care about how good the money is. Besides crypto, I don’t think there are many for many of the relevant people to make similar amounts of money on similar timeframes. Actually I think working at a lab early was an effective way to make money. A bunch of safety-concerned people for example have equity worth several millions to tens of millions, more than I think they could have easily earned elsewhere, and some are now billionaires on paper. And if AI has the transformative impact on the economy we expect, that could be worth way more (and it being worth more is correlated with it being needed more, so extra valuable); we are talking about the most valuable/powerful industry the world has ever known here, hard to beat that for making money. I don’t think that makes it okay to lead large AI labs, but for joining early, especially doing some capabilities work that doesn’t push the most risky capabilities along much, I don’t think it’s obvious.
I agree that there are various risks related to staying too long, rationalizing, being greedy, etc., and in most cases I wouldn’t advice a safety-concerned person to do capabilities. But I think you’re being substantially too intense about the risk of speeding up AI relative to the benefits of seeing what’s happening on the inside, which seem like they’ve already been very substantial
The Greg / Vasco bet reminded me of something: I went to buy a ceiling fan with a light in it recently. There was one on sale that happened to also tick all my boxes, joy! But the salesperson warned me “the light in this fan can’t be replaced and only has 10,000 hours in it. After that you’ll need a new fan. So you might not want to buy this one.” I chuckled internally and bought two of them, one for each room.
Has anyone seen an analysis that takes seriously the idea that people should eat some fruits, vegetables and legumes over others based on how much animal suffering they each cause?
I.e. don’t eat X fruit, eat Y one instead, because X fruit is [e.g.] harvested in Z way, which kills more [insert plausibly sentient creature].
Even though I’ve been in the AI Safety space for ~ 2 years, I can’t shake the feeling that every living thing dying painlessly in its sleep overnight (due to AI killing us) isn’t as bad as (i.e. is ‘better’ than) hundreds of millions of people living in poverty and/or hundreds of billions of animals being tortured.
This makes me suspicious of my motivations. I think I do the work partly because I kinda feel the loss of future generations, but mainly because AI Safety still feels so neglected (and my counter factual impact here is larger).
I don’t think s-risks play much of a role in this, although they have in the past (here, I define disempowerment due to AGI, or authoritarian use of AGI as s-risks).
Thanks for sharing. I suspect most of the hundreds of millions of people living in poverty would disagree with you, though, and would prefer not to painlessly die in their sleep tonight.
I think its possible we’re talking passed each other?
I don’t think he’s talking past you. His point seems that the vast majority of the hundreds of millions of people living in poverty both have net positive lives, and don’t want to die.
Even with a purely hedonistic outlook, it wouldn’t be better for their lives to end.
Unless you are not talking about the present, but a future far worse than today’s situation?
I’m saying that on some level it feels worse to me that 700 million people suffer in poverty than every single person dying painlessly in their sleep. Or that billions of animals are in torture factories. It sounds like I’m misunderstanding Jason’s point?
I would contend they are not “suffering” in poverty overall, because most of their lives are net positive. There may be many struggles and their lives are a lot harder than ours, but still better than not being alive at all.
I agree with you on the animals in torture factories, because their lives are probably net negative unlike the 700 million in poverty.
If AI actually does manage to kill us (which I doubt), It will not involve everybody dying painlessly in their sleep. That is an assumption of the “FOOM to god AI with no warning” model, which bears no resemblance to reality.
The technology to kill everyone on earth in their sleep instantaneously does not exist now, and will not exist in the near-future, even if AGI is invented. Killing everyone in their sleep is orders of magnitude more difficult than killing everyone awake, so why on earth would that be the default scenario?
I think you have a point with animals, but I don’t think the balance of human experience means that non-existence would be better than the status quo.
Will talks about this quite a lot in ch. 9 of WWOTF (“Will the future be good or bad?”). He writes:
And, of course, for people at least, things are getting better over time. I think animal suffering complicates this a lot.
Imagine running 100 simulations of humanity’s story. In every single one, the same pattern emerges: The moment we choose agriculture over hunting and gathering, we unknowingly start a countdown to our own extinction through AGI. If this were true, I think it suggests that our best chance at long-term survival is to stay as hunter-gatherers—and that what we call ‘progress’ is actually a kind of cosmic trap.
What makes the agriculture transition stand out as the “point of no return”? Agriculture was independently invented several times, so I’d expect the hunter-gatherer → agriculture transition to be more convergent than agriculture → AGI.
I mean, I’m mostly riffing here, but according to Claude: “Agriculture independently developed in approximately 7-12 major centers around the world.”
If we ran the simulation 100 times, would AGI appear in > 7-12 centres around the world? Maybe, I dunno.
Anyway, Happy Friday det!
If i found a place that raised cows that had predictably net positive lives, what would be the harm in eating beef from this farm?
I’ve been ostrovegan for ~ 7 years but open to changing my mind with new information.
This brief post might contain many relevant points: https://magnusvinding.com/2020/10/03/underappreciated-consequentialist-reasons/
thanks!
Here’s a list of non-animal welfare concerns about animal farming: https://forum.effectivealtruism.org/posts/rpBYejrhk7HQB6gJj/crispr-for-happier-farm-animals?commentId=hvCQ9kBvutrnkFm9h
thanks!
I expect that over the next couple of years GenAI products will continue to improve, but concerns about AI risk won’t. For some reason we can’t “feel” the improvements. Then (e.g. around 2026) we will have pretty capable digital agents and there will be a surge of concern (similar to when ChatGPT was released, maybe bigger), and then possibly another perceived (but not real) plateau.
The concerns in this case seem to be proportional to the actual real world impact: The jump from chatgpt and midjourney not existing to being available to millions was extremely noticeable: suddenly the average person could chat with a realistic bot and use it to cheat on their homework, or configure up a realistic looking image from a simple text prompt.
In contrast for the average person, not a lot has changed from 2022 to now. The chatbots give out less errors and the images are a lot better, but they aren’t able to accomplish significantly more actual tasks than they were in 2022. Companies are shoving AI tools into everything, but people are mostly ignoring them.
We seem to be boiling the frog, however I’m optimistic (perhaps naively) that GPT voice mode may wake some people up. “Just a chatbot” doesn’t carry quite the same weight when it’s actually speaking to you.
I really like this ad strategy from BlueDot. 5 stars.
90% of the message is still x / c-risk focussed, but by including “discrimination” and “loss of social connection” they’re clearly trying to either (or both);
create a big tent
nudge people that are in “ethics” but sympathetic to x / c-risk into x / c-risk
This feels like misleading advertising to me, particularly “Election interference” and “Loss of social connection”. Unless bluedot is doing some very different curriculum now and have a very different understanding of what alignment is. These bait-and-switch tactics might not be a good way to create a “big tent”.
Does c-risk stand for catastrophic risk?
haha yeah im trying to meme it into existence to make my life easier 😅
I’d like to quickly skill up on ‘working with people that have a kind of neurodivergence’.
This means;
understanding their unique challenges
how I can help them get the best from/for themselves
but I don’t want to sink more than two hours into this
Can anyone recommend some online content for me please?
Not sure if this type of concern has reached the meta yet, but if someone approached me asking for career advice, tossing up whether to apply for a job at a big AI lab, I would let them know that it could negatively affect their career prospects down the track because so many people now perceive such as a move as either morally wrong or just plain wrong-headed. And those perceptions might only increase over time. I am not making a claim here beyond this should be a career advice consideration.
Attention: AI Safety Enthusiasts in Wellington (and New Zealand)
I’m pleased to announce the launch of a brand new Facebook group dedicated to AI Safety in Wellington: AI Safety Wellington (AISW). This is your local hub for connecting with others passionate about ensuring a safe and beneficial future with artificial intelligence / reducing x-risk.To kick things off, we’re hosting a super casual meetup where you can:
Meet & Learn: Connect with Wellington’s AI Safety community.
Chat & Collaborate: Discuss career paths, upcoming events, and training opportunities.
Share & Explore: Whether you’re new to AI Safety or a seasoned expert, come learn from each other’s perspectives.
Expand the Network: Feel free to bring a friend who’s curious about AI Safety!
Join the group and RSVP for the meetup here
https://www.facebook.com/groups/500872105703980
Is there a Slack group or something similar for Founders of early-stage (EA) startups?
Something I’m confused about: what is the threshold that needs meeting for the majority of people in the EA community to say something like “it would be better if EAs didn’t work at OpenAI”?
Imagining the following hypothetical scenarios over 2024⁄25, I can’t predict confidently whether they’d individually cause that response within EA?
Ten-fifteen more OpenAI staff quit for varied and unclear reasons. No public info is gained outside of rumours
There is another board shakeup because senior leaders seem worried about Altman. Altman stays on
Superalignment team is disbanded
OpenAI doesn’t let UK or US AISI’s safety test GPT5/6 before release
There are strong rumours they’ve achieved weakly general AGI internally at end of 2025
A concrete prediction − 60% chance 80k doesn’t list any jobs at OpenAI in 2026.
I think it would be good if lots of EAs answered this twitter poll, so we could get a better sense for the communities views on the topic of Enlightenment / Awakening: https://twitter.com/SpencrGreenberg/status/1782525718586413085?ref_src=twsrc%5Egoogle%7Ctwcamp%5Eserp%7Ctwgr%5Etweet
The catchphrase I walk around with in my head regarding the optimal strategy for AI Safety is something like: Creating Superintelligent Artificial Agents* (SAA) without a worldwide referendum is ethically unjustifiable. Until a consensus is reached on whether to bring into existence such technology, a global moratorium is required (*we already have AGI).
I thought it might be useful to spell that out.
I recently discovered the idea of driving all blames into oneself, which immediately resonated with me. It is relatively hardcore; the kind of thing that would turn David Goggins into a Buddhist.
Gemini did a good job of summarising it:
This quote by Pema Chödron, a renowned Buddhist teacher, represents a core principle in some Buddhist traditions, particularly within Tibetan Buddhism. It’s called “taking full responsibility” or “taking self-blame” and can be a bit challenging to understand at first. Here’s a breakdown:
What it Doesn’t Mean:
Self-Flagellation: This practice isn’t about beating yourself up or dwelling on guilt.
Ignoring External Factors: It doesn’t deny the role of external circumstances in a situation.
What it Does Mean:
Owning Your Reaction: It’s about acknowledging how a situation makes you feel and taking responsibility for your own emotional response.
Shifting Focus: Instead of blaming others or dwelling on what you can’t control, you direct your attention to your own thoughts and reactions.
Breaking Negative Cycles: By understanding your own reactions, you can break free from negative thought patterns and choose a more skillful response.
Analogy:
Imagine a pebble thrown into a still pond. The pebble represents the external situation, and the ripples represent your emotional response. While you can’t control the pebble (the external situation), you can control the ripples (your reaction).
Benefits:
Reduced Suffering: By taking responsibility for your own reactions, you become less dependent on external circumstances for your happiness.
Increased Self-Awareness: It helps you understand your triggers and cultivate a more mindful response to situations.
Greater Personal Growth: By taking responsibility, you empower yourself to learn and grow from experiences.
Here are some additional points to consider:
This practice doesn’t mean excusing bad behavior. You can still hold others accountable while taking responsibility for your own reactions.
It’s a gradual process. Be patient with yourself as you learn to practice this approach.
What are some historical examples of a group (like AI Safety folk) getting something incredibly wrong about an incoming Technology? Bonus question: what led to that group getting it so wrong? Maybe there is something to learn here.
In the 90′s and 2000′s, many people such as Eric Drexler were extremely worried about nanotechnology and viewed it as an existential threat through the “gray goo” scenario. Yudkowsky predicted drexler style nanotech would occur by 2010, using very similar language to what he is currently saying about AGI.
It turned out they were all being absurdly overoptimistic about how soon the technology would arrive, and the whole drexlerite nanotech project flamed out by the end of the 2000′s and has pretty much not progressed since. I think a similar dynamic playing out with AGI is less likely, but still very plausible.
Do you have links to people being very worried about gray goo stuff?
(Also, the post you link to makes this clear, but this was a prediction from when Eliezer was a teenager, or just turned 20, which does not make for a particularly good comparison, IMO)
I hope you’re right. Thanks for the example, it seems like a good one.
This is probably a good exercise. I do want to point out a common bias about getting existential risks wrong. If someone was right about doomsday, we would not be here to discuss it. That is a huge survivorship bias. Even catestrophic events which lessen the number of people are going to be systemically underestimated. This phenomenon is the anthropic shadow which is relevant to an analysis like this.
Yeah, Case Studies as Research need to be treated very carefully (i.e. they can still be valuable exercises but the analyser needs to be aware of their weaknesses)
There were many predictions about AI and AGI in the past (maybe mostly last century) that were very wrong. I think I read about it in Superintelligence. A quick Google search shows this article which probably talks about that.
Thanks!
Cultured meat predictions were overly optimistic, although many of those predictions might have been companies hyping up their products to attract investors. There’s also probably a selection bias where the biggest cultured meat optimisits are the ones who become cultured meat experts and make predictions
https://pessimistsarchive.org/
I have heard rumours that an AI Safety documentary is being made. Separate to this, a good friend of mine is also seriously considering making one, but he isn’t “in” AI Safety. If you know who this first group is and can put me in touch with them, it might be worth getting across each others plans.
One of the seminal texts in marketing science is The Long and The Short of it by Field and Binet. They argue that for maximum effectiveness, marketing should aim to create two distinct work streams and results; immediate sales and longer term sales.
They argue the tactics that go into either are distinct (e.g. short term = create a sense of urgency, long term = match brand associations with category associations).
This feels like a good analogy for AI Safety Advocacy / Governance—keep talking about short term things people can buy now (in an Australian context that is the AI Safety Institute) while maintaining salience on the long term (x-risk, a global Pause, etc).
The question shouldn’t be either/or, rather, its about how much (%) effort should go into either at any point in time.
What would be the pros and cons of adding a semi-hidden-but-permanent Hot Takes section to the Forum? All of my takes are Hot and due to time constraints I would otherwise not post at all. Some would argue that someone like me should not post Hot Takes at all. Anyway, in true lazy fashion here is ChatGPT on the pros and cons:
Pros:
Encourages diverse perspectives and stimulates debate.
Can attract more engagement and interest from users.
Provides a platform for expressing unconventional or controversial ideas.
Fosters a culture of intellectual curiosity and open discourse within the community.
Cons:
May lead to increased polarization and conflict within the community.
Risk of spreading misinformation or poorly researched opinions.
Could divert attention from more rigorous and evidence-based discussions.
Potential for reputational damage if controversial opinions are associated with the forum.
This feels like it could just be a genre of Quick Takes that people may choose to post?
That’s an interesting one—I’m a fan of hot takes myself :D. I think “Quick takes” does the job on these though, even if the posts are a bit longer. I’m not sure we need another section. Maybe a “Hot takes” tab could be added to signify that the thought behind a take isn’t so deep?
Feels kinda obnoxious to write a quick take along the lines of “I’m thinking about writing a post on X, does anyone actually give a sh*t? Otherwise I won’t write it.”
I just wanted to check, since I can’t place my finger on why it feels obnoxious but it certainly does.
How impactful could non-dual meditation be for improving wellbeing?
Are there any EAs out there who have practiced non-dual meditation? Specifically, anything from the Dzogchen or Mahamudra traditions of Tibetan Buddhism?
More secular teachers would be Loch Kelly, Sam Harris, Michael Taft.
This has been a life changing experience for me and I’m wondering whether it could be a blind spot for EA.
I’d also just love to chat with someone else who has experienced non-duality / awakening through this form of meditation :)
I spent some time with Claude this morning trying to figure out why I find it cringe calling myself an EA (I never call myself an EA, even though many in EA would call me an EA).
The reason: calling myself “EA” feels cringe because it’s inherently a movement/community label—it always carries that social identity baggage with it, even when I’m just trying to describe my personal philosophical views.
I am happy to describe myself as a Buddhist or Utilitarian because I don’t think it does those things (at least, not within the broader community context I find myself in—Western, Online, Democratic, Australia, etc).
Reminds me of keeping your identity small by Paul Graham
[IMAGE] there is something about the lack overlap between these two audiences that makes me uneasy. WYD?
Prediction: In 6-12 months people are going to start leaving Deepmind and Anthropic for similar sounding reasons to those currently leaving OpenAI (50% likelihood).
> Surface level read of what is happening at OpenAI; employees are uncomfortable with specific safety policies.
> Deeper, more transferable, harder to solve problem; no person that is sufficiently well-meaning and close enough to the coal face at Big Labs can ever be reassured they’re doing the right thing continuing to work for a company whose mission is to build AGI.
Basically, this is less about “OpenAI is bad” and more “Making AGI is bad”.
Very quick thoughts on setting time aside for strategy, planning and implementation, since I’m into my 4th week of strategy development and experiencing intrusive thoughts about needing to hurry up on implementation;
I have a 52 week LTFF grant to do movement building in Australia (AI Safety)
I have set aside 4.5 weeks for research (interviews + landscape review + maybe survey) and strategy development (segmentation, targeting, positioning),
Then 1.5 weeks for planning (content, events, educational programs), during which I will get feedback from others on the plan and then iterate it.
This leaves me with 46⁄52 weeks to implement ruthlessly.
In conclusion, 6 weeks on strategy and planning seems about right. 2 weeks would have been too short, 10 weeks would have been too long, this porridge is juuuussttt rightttt.
keen for feedback from people in similar positions.
What would stop you from paying for an LLM? Take an extreme case; Sam Altman turns around tomorrow and says “We’re racing to AGI, I’m not going to worry about Safety at all.”
Would that stop you from throwing him $20 a month?
(I currently pay for Gemini)
Help clear something up for me: I am extremely confused (theoretically) how we can simultaneously have:
1. An Artificial Superintelligence
2. It be controlled by humans (therefore creating misuse of concentration of power issues)
My intuition is that once it reaches a particular level of power it will be uncontrollable. Unless people are saying that we can have models 100x more powerful than GPT4 without it having any agency??
I don’t think we need to explicitly alert the reader when we’ve received help from an LLM to write something (even if it wrote the bulk of the post). That’s it, my quickest ever Quick Take.
There have been multiple occasions where I’ve copy and pasted email threads into an LLM and asked it things like:
What is X person saying
What are the cruxes in this conversation?
Summarise this conversation
What are the key takeaways
What views are being missed from this conversation
I really want an email plugin that basically brute forces rationality INTO email conversations.
Tangentially—I wonder if LLMs can reliably convert peoples claims into a % through sentiment analysis? This would be useful for Forecasters I believe (and rationality in general)
It knows the concept of cruxes? I suppose that isn’t that surprising in retrospect.
Having a nondual Awakening was the second most important thing to happen to me (after my daughters birth). It has led to incredibly low levels of suffering and incredibly high levels of wellbeing. I write this because I think it is still incredibly under-appreciated and attainable for most people (maybe literally anyone).
There are traditions (Dzogchen, Zen, modern nonduality) where this shift in consciousness can be experienced simply by hearing the right combination of words and insights. As our understanding and tools for communicating these insights evolve, including through advances in AI, I believe this transformative experience could become accessible to many more people.
TLDR you can experience Awakening now, or, you can wait 3 years and an LLM will do it for you.
Is there an Animal Advocacy Careers (AAC) for AI Safety? If not, what is the most analogous org?
80,000 Hours.
I figured out why dudes like forecasting so much
Never thought I’d see the day of an EA / Aunty Donna crossover, but here we are. Thank you.
EA could do with another slice of pud
How bad does factory farming need to get before we WANT TO accelerate AI capabilities?
This article suggests that the DAILY “total painfulness of fish suffering is equivalent to the painfulness of one hundred fifty million human deaths” and that most of these deaths are via “slow suffocation, crushing, or live disemboweling”.
Let’s assume the creation of AGI had a >80% probability of leading to two outcomes for non-human animals: extinction or liberation.
How might we do the math on this?
At some point over the next 5 years virtually every person reading this will become a Pause advocate, the only question is whether that occurs now, in 2025, 2026, 2027, 2028 or 2029.
Flaming hot take: if you think Digital Sentience should be taken seriously but not Human Awakening / Enlightenment, then EA culture might have its hooks in a bit deep.
I like the “we decided to shut down our charity because it wasn’t very effective post” for the obvious reasons, but I wonder once you control for “is rich / comes from a rich family” and “doesn’t have family that depends on them” how that metric gets affected.
I think it is still good praise the decision in general, but unless I know the backgrounds of the people doing it I can’t heap too much praise on them.
If I’m understanding this concern correctly, it’s along the lines of: “they’re not making a financial sacrifice in shutting down, so it’s less praiseworthy than it otherwise would be”.
Just to clarify, charity founders (at least CE ones) take a pay cut to start their charity—they would earn more if working for other EA organizations as employees, and much more if in tech/finance/consulting/careers that typical of people with oxbridge/ivy/etc education levels. The financial sacrifice was already made when starting the charity, and if anything, quitting is actually better for you financially.
I’m confused about where you’re going with this. Why would the founders’ personal financial situation substantially affect how we viewed their decision?
In the absence of a poll feature, please use the agree/disagree function and the “changed my mind” emoji in this quick take to help me get a sense for EA’s views on a statement:
”Working on capabilities within a leading AI Lab makes someone a bad person”
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
Agree = strongly agree or somewhat agree
Disagree = strongly disagree or somewhat disagree
▲ reaction emoji = unsure / neither agree nor disagree
downvote = ~ this is a bad and divisive question
upvote = ~ this is a good question to be asking
If you listen to this, and eventually get to a point* where it doesn’t sound like gibberish (i.e. actually internalise), I guarantee your wellbeing will improve at least 10%.
*I imagine it will take on average about 6 months of work for a newbie.
I write this as I believe Awakening is still massively misunderstood by EA
https://deconstructingyourself.com/nonduality-and-awake-awareness-with-loch-kelly.html
I asked Midjourney-GPT-5 to visually represent the average brain of an Effective Altruist. Astounding results: https://www.instagram.com/reel/CtoGZDstA8k/?igshid=MzRlODBiNWFlZA==