I don’t think it’s odd at all. As the Bloomberg article notes, this was in response to the martyrdom of Saint Altman, when everyone thought the evil Effective Altruists were stealing/smashing OA for no reason and destroying humanity’s future (as well as the fortune of many of the individuals involved, to a degree few of them bothered to disclose) and/or turning it over to the Chinese commies. An internal memo decrying ‘Effective Altruism’ was far from the most extreme response at the time; but I doubt Gomez would write it today, if only because so much news has come out since then and it no longer looks like such a simple morality play. (For example, a lot of SV people were shocked to learn he had been fired from YC by Paul Graham et al for the same reason. That was a very closely held secret.)
gwern
Good report overall on tacit knowledge & biowarfare. This is relevant to the discussion over LLM risks: the Aum Shinrikyo chemist could make a lot of progress by reading papers and figuring out his problems as he went, but the bacteriologist couldn’t figure out his issues for what seems like what had been a viable plan to weaponize & mass-produce anthrax but where lack of feedback led it to fail. Which does sound like something that a superhumanly-knowledgeable (but not necessarily that intelligent) LLM could help a lot with simply by pattern-matching and making lists of suggestions for things that are to the human ‘unknown unknowns’.
If a crazy person wants to destroy the world with an AI-created bioweapon
Or, more concretely, nuclear weapons. Leaving aside regular full-scale nuclear war (which is censored from the graph for obvious reasons), this sort of graph will never show you something like Edward Teller’s “backyard bomb”, or a salted bomb. (Or any of the many other nuclear weapon concepts which never got developed, or were curtailed very early in deployment like neutron bombs, for historically-contingent reasons.)
There is, as far as I am aware, no serious scientific doubt that they are technically feasible: multi-gigaton bombs could be built or that salted bombs in relatively small quantities would render the earth uninhabitable to a substantial degree, for what are also modest expenditures as a percentage of GDP etc. It is just that there is no practical use of these weapons by normal, non-insane people. There is no use in setting an entire continent on fire, or in long-term radioactive poisoning of the same earth on which you presumably intend to live afterwards.
But you would be greatly mistaken if you concluded from historical data that these were impossible because there is nothing in the observed distribution anywhere close to those fatality rates.
(You can’t even make an argument from an Outside View of the sort that ‘there have been billions of humans and none have done this yet’, because nuclear bombs are still so historically new, and only a few nuclear powers were even in a position to consider whether to pursue these weapons or not—you don’t have k = billions, you have k < 10, maybe. And the fact that several of those pursued weapons like neutron bombs as far as they did, and that we know about so many concepts, is not encouraging.)
That was covered in my 2023-11-25 comment.
Hm, maybe it was common knowledge in some areas? I just always took him for being concerned. There’s not really any contradiction between being excited about your short-term work and worried about long-term risks. Fooling yourself about your current idea is an important skill for a researcher. (You ever hear the joke about Geoff Hinton? He suddenly solves how the brain works, at long last, and euphorically tells his daughter; she replies: “Oh Dad—not again!”)
Ilya has always been a doomer AFAICT, he was just loyal to Altman personally, who recruited him to OA. (I can tell you that when I spent a few hours chatting with him in… 2017 or something? a very long time ago, anyway—I don’t remember him dismissing the dangers or being pollyannaish.) ‘Superalignment’ didn’t come out of nowhere or surprise anyone about Ilya being in charge. Elon was… not loyal to Altman but appeared content to largely leave oversight of OA to Altman until he had one of his characteristic mood changes, got frustrated and tried to take over. In any case, he surely counts as a doomer by the time Zilis is being added to the board as his proxy. D’Angelo likewise seems to have consistently, in his few public quotes, been concerned about the danger.
A lot of people have indeed moved towards the ‘doomer’ pole but much of that has been timelines: AI doom in 2060 looks and feels a lot different from AI doom in 2027.
-
I haven’t seen any coverage of the double structure or Anthropic exit which suggests that Amodei helped think up or write the double structure. Certainly, the language they use around the Anthropic public benefit corporation indicates they all think, at least post-exit, that the OA double structure was a terrible idea (eg. see the end of this article).
-
You don’t know that. They seem to have often had near majorities, rather than being a token 1 or 2 board members.
By most standards, Karnofsky and Sutskever are ‘doomers’, and Zillis is likely a ‘doomer’ too as that is the whole premise of Neuralink and she was a Musk representative (which is why she was pushed out after Musk turned on OA publicly and began active hostilities like breaking contracts with OA). Hoffman’s views are hard to characterize, but he doesn’t seem to clearly come down as an anti-doomer or to be an Altman loyalist. (Which would be a good enough reason for Altman to push him out; and for a charismatic leader, neutralizing a co-founder is always useful, for the same reason no one would sell life insurance to an Old Bolshevik in Stalinist Russia.)
If I look at the best timeline of the board composition I’ve seen thus far, at a number of times post-2018, it looks like there was a ‘near majority’ or even outright majority. For example, 2020-12-31 has either a tie or an outright majority for either side depending on how you assume Sutskever & Hoffman (Sutskever?/Zilis/Karnofsky/D’Angelo/McCauley vs Hoffman? vs Altman/Brockman), and with the 2021-12-31 list the Altman faction needs to pick up every possible vote to match the existing 5 ‘EA’ faction (Zilis/Karnofsky/D’Angelo/McCauley/Toner vs Hurd?/Sutskever?/Hoffman? vs Brockman/Altman) although this has to be wrong because the board maxes out at 7 according to the bylaws so it’s unclear how exactly the plausible majorities evolved over time.
-
EDIT: this is going a bit viral, and it seems like many of the readers have missed key parts of the reporting. I wrote this as a reply to Wei Dai and a high-level summary for people who were already familiar with the details; I didn’t write this for people who were unfamiliar, and I’m not going to reference every single claim in it, as I have generally referenced them in my prior comments/tweets and explained the details & inferences there. If you are unaware of aspects like ‘Altman was trying to get Toner fired’ or pushing out Hoffman or how Slack was involved in Sutskever’s flip or why Sutskever flip-flopped back, still think Q* matters, haven’t noticed the emphasis put on the promised independent report, haven’t read the old NYer Altman profile or Labenz’s redteam experience etc., it may be helpful to catch up by looking at other sources; my comments have been primarily on LW since I’m not a heavy EAF user, plus my usual excerpts.
Or even “EA had a pretty weak hand throughout and played it as well as can be reasonably expected”?
It was a pretty weak hand. There is this pervasive attitude that Sam Altman could have been dispensed with easily by the OA Board if it had been more competent; this strange implicit assumption that Altman is some johnny-come-lately where the Board screwed up by hiring him. Commenters seem to ignore the long history here—if anything, it was he who screwed up by hiring the Board!
Altman co-founded OA. He was the face in initial coverage and 1 of 2 board members (with Musk). He was a major funder of it. Even Elon Musk’s main funding of OA was through an Altman vehicle. He kicked out Musk when Musk decided he needed to be in charge of OA. Open Philanthropy (OP) only had that board seat and made a donation because Altman invited them to, and he could personally have covered the $30m or whatever OP donated for the seat; and no one cared or noticed when OP let the arrange lapse after the initial 3 years. (I had to contact OP to confirm this when someone doubted that the seat was no longer controlled by OP.) He thought up, drafted, and oversaw the entire for-profit thing in the first place, including all provisions related to board control. He voted for all the board members, filling it back up from when it was just him (& Greg Brockman at one point IIRC). He then oversaw and drafted all of the contracts with MS and others, while running the for-profit and eschewing equity in the for-profit. He designed the board to be able to fire the CEO because, to quote him, “the board should be able to fire me”. He interviewed every person OA hired, and used his networks to recruit for OA. And so on and so forth
Credit where credit is due—Altman may not have believed the scaling hypothesis like Dario Amodei, may not have invented PPO like John Schulman, may not have worked on DL from the start like Ilya Sutskever, may not have created GPT like Alec Radford, may not have written & optimized any code like Brockman’s—but the 2023 OA organization is fundamentally his work.
The question isn’t, “how could EAers* have ever let Altman take over OA and possibly kick them out”, but entirely the opposite: “how did EAers ever get any control of OA, such that they could even possibly kick out Altman?” Why was this even a thing given that OA was, to such an extent, an Altman creation?
The answer is: “because he gave it to them.” Altman freely and voluntarily handed it over to them.
So you have an answer right there to why the Board was willing to assume Altman’s good faith for so long, despite everyone clamoring to explain how (in hindsight) it was so obvious that the Board should always have been at war with Altman and regarding him as an evil schemer out to get them. But that’s an insane way for them to think! Why would he undermine the Board or try to take it over, when he was the Board at one point, and when he made and designed it in the first place? Why would he be money-hungry when he refused all the equity that he could so easily have taken—and in fact, various partner organizations wanted him to have in order to ensure he had ‘skin in the game’? Why would he go out of his way to make the double non-profit with such onerous & unprecedented terms for any investors, which caused a lot of difficulties in getting investment and Microsoft had to think seriously about, if he just didn’t genuinely care or believe any of that? Why any of this?
(None of that was a requirement, or even that useful to OA for-profit. Other double systems like Mozilla or Hershey don’t have such terms, they’re just normal corporations with a lot of shares owned by a non-profit, is all. OA for-profit could’ve been the same way. Certainly, if all of this was for PR reasons or some insidious decade-long scheme of Altman to ‘greenwash’ OA, it was a spectacular failure—nothing has occasioned more confusion and bad PR for OA than the double structure or capped-profit. See, for example, my shortly-before-the-firing Twitter argument with well-known AI researcher Delip Rao who repeatedly stated & doubled down on the claim that the OA non-profit legally owned the OA for-profit was not just factually wrong but misinformation. He helpfully linked to a page about political misinformation & propaganda campaigns online in case I had any doubt about what the term ‘misinformation’ meant.)
What happened is, broadly: ‘Altman made the OA non/for-profits and gifted most of it to EA with the best of intentions, but then it went so well & was going to make so much money that he had giver’s remorse, changed his mind, and tried to quietly take it back; but he had to do it by hook or by crook, because the legal terms said clearly “no takesie backsies”’. Altman was all for EA and AI safety and an all-powerful nonprofit board being able to fire him, and was sincere about all that, until OA & the scaling hypothesis succeeded beyond his wildest dreams†, and he discovered it was inconvenient for him and convinced himself that the noble mission now required him to be in absolute control, never mind what restraints on himself he set up years ago—he now understands how well-intentioned but misguided he was and how he should have trusted himself more. (Insert Garfield meme here.)
No wonder the board found it hard to believe! No wonder it took so long to realize Altman had flipped on them, and it seemed Sutskever needed Slack screenshots showing Altman blatantly lying to them about Toner before he finally, reluctantly, flipped. The Altman you need to distrust & assume bad faith of & need to be paranoid about stealing your power is also usually an Altman who never gave you any power in the first place! I’m still kinda baffled by it, personally.
He concealed this change of heart from everyone, including the board, gradually began trying to unwind it, overplayed his hand at one point—and here we are.
So, what could the EA faction of the board have done? …Not much, really. They only ever had the power that Altman gave them in the first place.
* I don’t really agree with this framing of Sutskever/Toner/McCauley/D’Angelo as “EA”, but for the sake of argument, I’ll go with this labeling.
† Please try to cast your mind back to when Altman et al would be planning all this in 2018-2019, with OA rapidly running out of cash after the mercurial Musk’s unexpected-but-inevitable betrayal, its DRL projects like OA5 remarkable research successes but commercially worthless, and just some interesting results like GPT-1 and then GPT-2-small coming out of their unsupervised learning backwater from Alec Radford tinkering around with RNNs and then these new ‘Transformer’ things. The idea that OA might somehow be worth over ninety billion dollars, yes, that’s ‘billion’ with a ‘b’ in scarcely 3 years would have been insane, absolutely insane, not a single person in the AI world would have taken you seriously if you had suggested that and if you emailed any of them asking about how plausible that was, they would have added an email filter to send your future emails to the trash bin. It is very easy to be sincerely full of the best intentions and discuss how to structure your double corporation to deal with windfalls like growing to a $1000b market cap when no one really expects that to happen, certainly not in the immediate future… Thus, no one is also sitting around going, ‘well wait, we required the board to not own equity, but if the company is worth even fraction of our long-term targets, and it’s recruiting with stock options like usual, then each employee is going to have, like, $10m or even $100m of pseudo-equity in the OA for-profit. That seems… problematic. Do we need to do something about it?’
By the time Musk (and Altman et al) was starting OA, it was in response to Page buying Hassabis. So there is no real contradiction here between being spurred by Page’s attitude and treating Hassabis as the specific enemy. It’s not like Page was personally overseeing DeepMind (or Google Brain) research projects, and Page quasi-retired about a year after the DM purchase anyway (and about half a year before OA officially became a thing).
The discussion of the Abbey, er, I mean, ‘castle’, has been amusing for showing how much people are willing to sound off on topics from a single obviously-untrustworthy photograph. Have you ever seen a photograph of the interior or a layout? No, you merely see the single aerial real estate brochure shot using a telephoto zoom lenses framed as flatteringly as possible to include stuff that isn’t even the Abbey—like that turreted ‘castle’ you see in the photo up above isn’t even part of the Abbey—because that’s an active church, All Saints Church!* (Really, apply some critical thinking here: you think some manor house one can buy will just have a bunch of visible graves in it...?)
Let me ask something: how many of the people debating the Abbey on this page have been there? I don’t see anyone directly addressing the core claim of ‘luxury’, so I will.
I was there for an AI workshop earlier this year in Spring and stayed for 2 or 3 days, so let me tell you about the ‘luxury’ of the ‘EA castle’: it’s a big, empty, cold, stone box, with an awkward layout. (People kept getting lost trying to find the bathroom or a specific room.) Most of the furnishings were gone. Much of the layout you can see in Google Maps was nonfunctional, and several wings were off-limits or defunct, so in practice it was maybe a quarter of the size you’d expect from the Google Maps overview. There were clearly extensive needs for repair and remodeling of a lot of ancient construction, and most of the gardens are abandoned as too expensive to maintain. It is, as a real estate agent might say, very ‘historical’ and a ‘good fixer-upper’.
The kitchen area is pretty nice, but the part of the Abbey I admired most, from the standpoint of ‘luxury’, was the view of the neighboring farmer’s field. (It was extraordinarily green and verdant and picturesque, truly exemplifying “green and pleasant land”. I tried to take some photos on my phone, but they never capture the wetness & coloring.)
Otherwise, the best efforts of the hardworking staff at the workshop notwithstanding—and I’m trying not to make this sound like an insult—I would rate the level of ‘luxury’ as roughly ‘student hostel’ level. (Which is fully acceptable to me, but anyone expecting ‘luxury’ or the elite lifestyle of the Western nomenklatura is going to be disappointed. Windsor or Balmoral or a 5-star hotel, this is not.) Indeed, I’m not sure how the place could be in much rougher shape while still being an acceptable setting for a conference. (Once you’re down to a big mattress in an empty room, it’s hard to go down further without, like, removing electricity and indoor plumbing.)
The virtue of the Abbey is that it can be relatively easily reached from London/the rest of the world by simple public transit routes that even a first-time foreigner can navigate successfully, and contain a decent number of people without paying extortionate Oxford hotel rates or forcing people to waste hours a day going back & forth between their own lodgings creating lots of overhead in coordination. (“Oh, you should talk to Jack about that! oh, he just called an Uber for his hotel. Never mind.”)
Buying it seems entirely reasonable to me assuming adequate utilization in terms of hosting events. (Which may or may not be the case, but no one here or elsewhere is even attempting to make it.) Nor do I see why any sort of public discussion would be so important and so scandalous to not have, because this is a subject on which any sort of ‘public discussion’ would be pointless—random Internet commenters don’t have a better idea of the FHI/EA event calendar or constraints of Oxford hotel booking than the people who were making the decisions here.
* I don’t know if technically it sits on the Abbey parcel or what, because England has lots of weird situations like that, but EA and EA visitors are obviously not getting any good or ‘luxury’ out of an active church regardless of its de jure status (we made no use of it in any way I saw), and including it in the image is misleading in an ordinary realtor sort of way.
- 13 Dec 2023 11:03 UTC; 32 points) 's comment on Nonlinear’s Evidence: Debunking False and Misleading Claims by (
Current reporting is that ‘EAs out of the board’ (starting with expelling Toner for ‘criticizing’ OA) was the explicit description/goal told to Sutskever shortly before, with reasons like being to avoid ‘being painted in the press as “a bunch of effective altruists,” as one of them put it’.
Whatever the nature of Q*, there is not much evidence that it could have prompted the Altman firing. It’s not clear why some very early preliminary results about a Q* as described would prompt a firing nor why the firing would be so abrupt or when it was if the research happened months ago (and Altman was alluding to it publicly weeks ago), while Sutskever’s involvement & the exact timing of the firing appear to be adequately explained by other issues.
As there is still nothing leaking or confirming Q*, I’m increasingly skeptical of its relevance—for something supposedly being demoed and discussed company-wide, if it was so cosmically important or so safety-relevant, you’d think there’d be more meat on the rumor bones by now, and there wouldn’t be issues like denying that Murati confirmed the rumors as opposed to merely described the rumors (which is the sort of garbling that is in line with past leaks like the initial misdescriptions of Sutskever refusing to give examples of Altman’s deceptive behavior). This increasingly sounds like a real system (maybe) which has been seized on by hype and yoked to a scandal it’s mostly uninvolved with. (Maybe related to issues of candor, but not ‘the’ reason.)
Discussing Q* seems like a big distraction and waste of time, IMO, when there are better-reported things to discuss (like Altman trying to fire Helen Toner from the OA board).
I’m confused why people are broadcasting such a tiny morsel of news so much, and it’s getting so much play on places like Marginal Revolution, while other parts of the recent OA drama don’t seem to rate a mention. Nobody seemed to care even 1% as much about the prior OA research like GPT-f or process evaluation...
Maybe? I can’t easily appreciate such a usecase because I always want to save any excerpts I find worth excerpting. Are there a lot of people who want that? If that’s the idea, I guess the “About Highlights” dialogue needs a bit of documentation to explain the intended (and unintended) uses. At least, anyone who doesn’t realize that the annotations are ephemeral, because they aren’t enough of a web dev to understand ‘What you save is stored only on your specific browser locally’ is as much of a bug as it is a feature, is in for a bad time when their annotations inevitably get deleted...
I like it overall.
But I have a lot of questions about the ‘highlight’ feature: aside from the many teething problems Said has already documented, which bugs doubtless will be fixed, I don’t understand what the usecase is, compared to other web annotation systems like Hypothesis—so it stores arbitrary ranges of text, saving them to, I assume, dangerously ephemeral browser LocalStorage where they will be unpredictably erased in a few hours / days / weeks, and unavailable on any other device presumably. Why do I want this? They aren’t clipped to something like Evernote, they aren’t synced to my phone, they aren’t saved anywhere safe, they aren’t posted to social media or visible to anyone else in any way… Is the idea to highlight various parts and then manually copy-paste them?
I agree. I’ve already started to develop a bit of an instinctive ‘ugh’ reaction to random illustrations, even ones without obvious generative model tell-tales or the DALL-E watermark.
It’s comparable to how you feel when you notice that little ‘© Getty Images’ or a Memphis style image, and realize your time & (mental) bandwidth was wasted by the image equivalent of an emoji. It’s not that they look bad, necessarily, but they increasingly signify ‘cheap’ and ‘tacky’. (After all, if this monkey pox image can be generated by a prompt of 12 redundant words, then that’s another of saying that the image is worth far less than a thousand words—it’s worth less than 12...)
It gets worse if you can spot any flaws or they do look bad overall: “so you couldn’t even take another minute to generate a variant without blatant artifacting? That’s how slovenly and careless your work is, and how little you respect your readers?”
For those wondering why we needed a stylish magazine for provocative rationalist/EA nonfiction when Works In Progress is pretty good too, Scott Alexander says
Works In Progress is a Progress Studies magazine, I’m sure these two movements look exactly the same to everyone on the outside, but we’re very invested in the differences between them.
What disease would you seek FDA approval for? “I sleep more than 4 hours a day” is not a recognized disease under the status quo. (There is the catch-all of ‘hypersomnia’, but things like sleep apnea or neurodegenerative disorders or damage to clock-keeping neurons would not plausibly be treated by some sort of knockout-mimicking drug.)
One downside you don’t mention: having a Wikipedia article can be a liability when editors are malicious, for all the reasons it is a benefit when it is high-quality like its popularity and mutability. A zealous attacker or deletionist destroying your article for jollies is bad, but at least it merely undoes your contribution and you can mirror it; an article being hijacked (which is what a real attacker will do) can cause you much more damage than you would ever have gained as it creates a new reality which will echo everywhere.
My (unfortunately very longstanding) example of this is the WP article on cryonics: you will note that the article is surprisingly short for a topic on which so much could be said and reads like it’s been barely touched in half a decade. Strikingly, while having almost no room for any information on minor topics like how cryonics works or how current cryonics orgs operate or the background on why it should be possible in principle or remarkable research findings like the progress on bringing pigs back from the dead, instead, the introduction, and an entire section, harp on how corporations go bankrupt and it is unlikely that a corporation today will be around in a century and how ancient pre-1973 cryonics companies have all gone bankrupt and so on. These claims are mostly true, but you will then search the article in vain for any mention that the myriad of cryonics bankruptcies alluded to is like 2 or 3 companies, that cryonics for the past 50 years isn’t done solely by corporations precisely because of that when it became apparent that cryonics was going to need to be a long-term thing & families couldn’t be trusted to pay, they are structured as trusts (the one throwaway comma mentioning trusts is actively misleading by implying that they are optional and unusual, rather than the status quo), and that there have been few or no bankruptcies or known defrostings since. All attempts to get any of this basic information into the article is blocked by editors. Anyone who comes away with an extremely negative opinion of cryonics can’t be blamed when so much is omitted to put it in the worst possible light. You would have to be favorably disposed to cryonics already to be reading this article and critically thinking to yourself, “did cryonicists really learn nothing from the failures? how do cryonicists deal with these criticisms when they are so obvious, it doesn’t seem to say? if the cryonics orgs go bankrupt so often, why doesn’t it name any of the many bankruptcies in the 49 years between 1973 and 2022, and how are any of these orgs still around?” etc.
More recently, the Scott Alexander/NYT fuss: long-time WP editor & ex-LWer David Gerard finally got himself outright topic-banned from the SA WP article when he overreached by boasting on Twitter how he was feeding claims to the NYT journalist so the journalist could print them in the article in some form and Gerard could then cite them in the WP article (and safe to say, any of the context or butt-covering caveats in the article version would be sanded away and simplified in the WP version to the most damaging possible version, which would then be defended as obviously relevant and clearly WP:V to an unimpeachable WP:RS). Gerard and activists also have a similar ‘citogenesis’ game going with Rational Wiki and friendly academics laundering into WP proper: make allegations there, watch them eventually show up in a publication of some sort, however tangential, and now you can add to the target article “X has been described as a [extremist / white supremacist / racist / fringe figure / crackpot] by [the SPLC / extremism researchers / the NYT / experts / the WHO]
<ref></ref>
”. Which will be true—there will in fact be a sentence, maybe even two or three about it in the ref. And there the negative statements will stay forever if they have anything to say about it (which they do), while everything else positive in the article dies the death of a thousand cuts. This can then be extended: do they have publications in some periodicals? Well, extremist periodicals are hardly WP:RSes now are they and shouldn’t be cited (WP:NAZI)… Scott’s WP article may not be too bad right now, but one is unlikely to be so lucky to get such crystal-clear admissions of bad faith editing, a large audience of interested editors going beyond the usual suspects of self-selected activist-editors who are unwilling to make excuses for the behavior, and despite all that, who knows how the article will read a year or a decade from now?
If Mark is in such a situation (which he was not, and I knew he was not), then the real culprit is whoever implemented such a completely broken and utterly unfixable ‘anonymous’ comment, and himself for being a security researcher and yet believing that retroactively making comments ‘anonymous’ on a publicly-scrape-able website would protect him against nation-state actors when anonymity was neither the goal nor a documented promise of the account deletion feature he was abusing and then crying ‘dox!’ about it not doing what it wasn’t supposed to do and didn’t do.