AI Safety NewsletterCenter for AI SafetyMay 16, 2023, 3:22 PMAI Safety Newsletter #1 [CAIS Linkpost]AkashApr 10, 2023, 8:18 PM38 points0 comments1 min readEA linkAI Safety Newsletter #2: ChaosGPT, Natural Selection, and AI Safety in the MediaOliver ZApr 18, 2023, 6:36 PM56 points1 comment4 min readEA link(newsletter.safe.ai)AI Safety Newsletter #3: AI policy proposals and a new challenger approachesOliver ZApr 25, 2023, 4:15 PM35 points1 comment4 min readEA link(newsletter.safe.ai)AI Safety Newsletter #4: AI and Cybersecurity, Persuasive AIs, Weaponization, and Geoffrey Hinton talks AI risksCenter for AI SafetyMay 2, 2023, 4:51 PM35 points2 comments5 min readEA link(newsletter.safe.ai)AI Safety Newsletter #5: Geoffrey Hinton speaks out on AI risk, the White House meets with AI labs, and Trojan attacks on language modelsCenter for AI SafetyMay 9, 2023, 3:26 PM60 points0 comments4 min readEA link(newsletter.safe.ai)AI Safety Newsletter #6: Examples of AI safety progress, Yoshua Bengio proposes a ban on AI agents, and lessons from nuclear arms controlCenter for AI SafetyMay 16, 2023, 3:14 PM32 points1 comment6 min readEA link(newsletter.safe.ai)AI Safety Newsletter #7: Disinformation, Governance Recommendations for AI labs, and Senate Hearings on AICenter for AI SafetyMay 23, 2023, 9:42 PM23 points0 comments6 min readEA link(newsletter.safe.ai)AI Safety Newsletter #8: Rogue AIs, how to screen for AI risks, and grants for research on democratic governance of AICenter for AI SafetyMay 30, 2023, 11:44 AM16 points3 comments6 min readEA link(newsletter.safe.ai)AISN #9: Statement on Extinction Risks, Competitive Pressures, and When Will AI Reach Human-Level?Center for AI SafetyJun 6, 2023, 3:56 PM12 points2 comments7 min readEA link(newsletter.safe.ai)AISN #12: Policy Proposals from NTIA’s Request for Comment and Reconsidering Instrumental ConvergenceCenter for AI SafetyJun 27, 2023, 3:25 PM30 points3 comments7 min readEA link(newsletter.safe.ai)AISN #13: An interdisciplinary perspective on AI proxy failures, new competitors to ChatGPT, and prompting language models to misbehaveCenter for AI SafetyJul 5, 2023, 3:33 PM25 points0 comments9 min readEA link(newsletter.safe.ai)AISN#14: OpenAI’s ‘Superalignment’ team, Musk’s xAI launches, and developments in military AI useCenter for AI SafetyJul 12, 2023, 4:58 PM26 points0 comments4 min readEA link(newsletter.safe.ai)AISN #16: White House Secures Voluntary Commitments from Leading AI Labs and Lessons from OppenheimerCenter for AI SafetyJul 25, 2023, 4:45 PM7 points0 comments6 min readEA link(newsletter.safe.ai)AISN #17: Automatically Circumventing LLM Guardrails, the Frontier Model Forum, and Senate Hearing on AI OversightCenter for AI SafetyAug 1, 2023, 3:24 PM15 points0 comments8 min readEA linkAISN #18: Challenges of Reinforcement Learning from Human Feedback, Microsoft’s Security Breach, and Conceptual Research on AI SafetyCenter for AI SafetyAug 8, 2023, 3:52 PM12 points0 comments5 min readEA link(newsletter.safe.ai)AISN #20: LLM Proliferation, AI Deception, and Continuing Drivers of AI CapabilitiesCenter for AI SafetyAug 29, 2023, 3:03 PM12 points0 comments8 min readEA link(newsletter.safe.ai)AISN #21: Google DeepMind’s GPT-4 Competitor, Military Investments in Autonomous Drones, The UK AI Safety Summit, and Case Studies in AI PolicyCenter for AI SafetySep 5, 2023, 2:59 PM13 points0 comments5 min readEA link(newsletter.safe.ai)AISN #22: The Landscape of US AI Legislation - Hearings, Frameworks, Bills, and LawsCenter for AI SafetySep 19, 2023, 2:43 PM15 points1 comment5 min readEA link(newsletter.safe.ai)AISN #23: New OpenAI Models, News from Anthropic, and Representation EngineeringCenter for AI SafetyOct 4, 2023, 5:10 PM7 points0 comments5 min readEA link(newsletter.safe.ai)AISN #24: Kissinger Urges US-China Cooperation on AI, China’s New AI Law, US Export Controls, International Institutions, and Open Source AICenter for AI SafetyOct 18, 2023, 5:03 PM16 points1 comment6 min readEA link(newsletter.safe.ai)AISN #25: White House Executive Order on AI, UK AI Safety Summit, and Progress on Voluntary Evaluations of AI RisksCenter for AI SafetyOct 31, 2023, 7:24 PM21 points0 comments6 min readEA link(newsletter.safe.ai)AISN #26: National Institutions for AI Safety, Results From the UK Summit, and New Releases From OpenAI and xAICenter for AI SafetyNov 15, 2023, 4:03 PM11 points0 comments6 min readEA link(newsletter.safe.ai)AISN #27: Defensive Accelerationism, A Retrospective On The OpenAI Board Saga, And A New AI Bill From Senators Thune And KlobucharCenter for AI SafetyDec 7, 2023, 3:57 PM10 points0 comments6 min readEA link(newsletter.safe.ai)AISN #28: Center for AI Safety 2023 Year in ReviewCenter for AI SafetyDec 23, 2023, 9:31 PM17 points1 comment5 min readEA link(newsletter.safe.ai)AISN #29: Progress on the EU AI Act Plus, the NY Times sues OpenAI for Copyright Infringement, and Congressional Questions about Research Standards in AI SafetyCenter for AI SafetyJan 4, 2024, 4:03 PM5 points0 comments6 min readEA link(newsletter.safe.ai)AISN #30: Investments in Compute and Military AI Plus, Japan and Singapore’s National AI Safety InstitutesCenter for AI SafetyJan 24, 2024, 7:38 PM7 points1 comment6 min readEA link(newsletter.safe.ai)AISN #31: A New AI Policy Bill in California Plus, Precedents for AI Governance and The EU AI OfficeCenter for AI SafetyFeb 21, 2024, 9:55 PM27 points0 comments6 min readEA link(newsletter.safe.ai)AISN #32: Measuring and Reducing Hazardous Knowledge in LLMs Plus, Forecasting the Future with LLMs, and Regulatory MarketsCenter for AI SafetyMar 7, 2024, 4:37 PM15 points2 comments8 min readEA link(newsletter.safe.ai)AISN #33: Reassessing AI and Biorisk Plus, Consolidation in the Corporate AI Landscape, and National Investments in AICenter for AI SafetyApr 12, 2024, 4:11 PM19 points0 comments9 min readEA link(newsletter.safe.ai)AISN #34: New Military AI Systems Plus, AI Labs Fail to Uphold Voluntary Commitments to UK AI Safety Institute, and New AI Policy Proposals in the US SenateCenter for AI SafetyMay 2, 2024, 4:12 PM21 points5 comments8 min readEA link(newsletter.safe.ai)AISN #35: Lobbying on AI Regulation Plus, New Models from OpenAI and Google, and Legal Regimes for Training on Copyrighted DataCenter for AI SafetyMay 16, 2024, 2:26 PM14 points0 comments6 min readEA link(newsletter.safe.ai)AISN #36: Voluntary Commitments are Insufficient Plus, a Senate AI Policy Roadmap, and Chapter 1: An Overview of Catastrophic RisksCenter for AI SafetyMay 30, 2024, 6:23 PM6 points0 comments5 min readEA link(newsletter.safe.ai)AI Safety Newsletter #37: US Launches Antitrust Investigations Plus, recent criticisms of OpenAI and Anthropic, and a summary of Situational AwarenessCenter for AI SafetyJun 18, 2024, 6:08 PM15 points0 comments5 min readEA link(newsletter.safe.ai)AISN #38: Supreme Court Decision Could Limit Federal Ability to Regulate AI Plus, “Circuit Breakers” for AI systems, and updates on China’s AI industryCenter for AI SafetyJul 9, 2024, 7:29 PM8 points0 comments5 min readEA link(newsletter.safe.ai)AI Safety Newsletter #40: California AI Legislation Plus, NVIDIA Delays Chip Production, and Do AI Safety Benchmarks Actually Measure Safety?Center for AI SafetyAug 21, 2024, 6:10 PM17 points0 comments6 min readEA link(newsletter.safe.ai)AI Safety Newsletter #39: Implications of a Trump Administration for AI Policy Plus, Safety EngineeringCenter for AI SafetyJul 29, 2024, 5:48 PM6 points0 comments6 min readEA link(newsletter.safe.ai)AI Safety Newsletter #41: The Next Generation of Compute Scale Plus, Ranking Models by Susceptibility to Jailbreaking, and Machine EthicsCenter for AI SafetySep 11, 2024, 7:11 PM12 points0 comments5 min readEA link(newsletter.safe.ai)AI Safety Newsletter #42: Newsom Vetoes SB 1047 Plus, OpenAI’s o1, and AI Governance SummaryCenter for AI SafetyOct 1, 2024, 8:33 PM10 points0 comments6 min readEA link(newsletter.safe.ai)AI Safety Newsletter #43: White House Issues First National Security Memo on AI Plus, AI and Job Displacement, and AI Takes Over the NobelsCenter for AI SafetyOct 28, 2024, 4:02 PM6 points0 comments6 min readEA link(newsletter.safe.ai)AISN #44: The Trump Circle on AI Safety Plus, Chinese researchers used Llama to create a military tool for the PLA, a Google AI system discovered a zero-day cybersecurity vulnerability, and Complex Systems Center for AI SafetyNov 19, 2024, 4:36 PM11 points0 comments5 min readEA link(newsletter.safe.ai)