Narrations of the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. This podcast also contains narrations of some of our publications. ABOUT US The Center for AI Safety (CAIS) is a San Francisco-based research and field-building nonprofit. We believe that artificial intelligence has the potential to profoundly benefit the world, provided that we can develop and use it safely. However, in contrast to the dramatic p ...
…
continue reading
The Into AI Safety podcast aims to make it easier for everyone, regardless of background, to get meaningfully involved with the conversations surrounding the rules and regulations which should govern the research, development, deployment, and use of the technologies encompassed by the term "artificial intelligence" or "AI" For better formatted show notes, additional resources, and more, go to https://into-ai-safety.github.io For even more content and community engagement, head over to my Pat ...
…
continue reading

1
AISN #48: Utility Engineering and EnigmaEval
8:56
8:56
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
8:56Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. In this newsletter, we explore two recent papers from CAIS. We’d also like to highlight that CAIS is hiring for editorial and writin…
…
continue reading
Plus, State-Sponsored AI Cyberattacks. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. Reasoning Models DeepSeek-R1 has been one of the most significant model releases since ChatGPT. After its release, the DeepSeek's app quickly rose to the top of Apple's most downloaded chart and NVIDIA saw a 17% stock decline. In this st…
…
continue reading
Plus, Humanity's Last Exam, and the AI Safety, Ethics, and Society Course. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. The Transition The transition from the Biden to Trump administrations saw a flurry of executive activity on AI policy, with Biden signing several last-minute executive orders and Trump revoking Biden's…
…
continue reading

1
AISN #45: Center for AI Safety 2024 Year in Review
11:31
11:31
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
11:31As 2024 draws to a close, we want to thank you for your continued support for AI safety and review what we’ve been able to accomplish. In this special-edition newsletter, we highlight some of our most important projects from the year. The mission of the Center for AI Safety is to reduce societal-scale risks from AI. We focus on three pillars of wor…
…
continue reading

1
AISN #44: The Trump Circle on AI Safety
11:22
11:22
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
11:22Plus, Chinese researchers used Llama to create a military tool for the PLA, a Google AI system discovered a zero-day cybersecurity vulnerability, and Complex Systems. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. The Trump Circle on AI Safety The incoming Trump administration is likely to significantly alter the US gover…
…
continue reading

1
AISN #43: White House Issues First National Security Memo on AI
14:55
14:55
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
14:55Plus, AI and Job Displacement, and AI Takes Over the Nobels. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. White House Issues First National Security Memo on AI On October 24, 2024, the White House issued the first National Security Memorandum (NSM) on Artificial Intelligence, accompanied by a Framework to Advance AI Gov…
…
continue reading
Plus, OpenAI's o1, and AI Governance Summary. Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. Newsom Vetoes SB 1047 On Sunday, Governor Newsom vetoed California's Senate Bill 1047 …
…
continue reading

1
AISN #41: The Next Generation of Compute Scale
11:59
11:59
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
11:59Plus, Ranking Models by Susceptibility to Jailbreaking, and Machine Ethics. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. The Next Generation of Compute Scale AI development is on the cusp of a dramatic expansion in compute scale. Recent developments across multiple fronts—from chip manufacturing to power infrastructure—…
…
continue reading
Plus, NVIDIA Delays Chip Production, and Do AI Safety Benchmarks Actually Measure Safety?. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. SB 1047, the Most-Discussed California AI Legislation California's Senate Bill 1047 has sparked discussion over AI regulation. While state bills often fly under the radar, SB 1047 has g…
…
continue reading

1
AISN #39: Implications of a Trump Administration for AI Policy
12:00
12:00
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
12:00Plus, Safety Engineering Overview. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. Implications of a Trump administration for AI policy Trump named Ohio Senator J.D. Vance—an AI regulation skeptic—as his pick for vice president. This choice sheds light on the AI policy landscape under a future Trump administration. In this…
…
continue reading

1
AISN #38: Supreme Court Decision Could Limit Federal Ability to Regulate AI
10:31
10:31
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
10:31Plus, “Circuit Breakers” for AI systems, and updates on China's AI industry. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. Supreme Court Decision Could Limit Federal Ability to Regulate AI In a recent decision, the Supreme Court overruled the 1984 precedent Chevron v. Natural Resources Defence Council. In this story, we …
…
continue reading

1
AISN #37: US Launches Antitrust Investigations
11:02
11:02
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
11:02US Launches Antitrust Investigations The U.S. Government has launched antitrust investigations into Nvidia, OpenAI, and Microsoft. The U.S. Department of Justice (DOJ) and Federal Trade Commission (FTC) have agreed to investigate potential antitrust violations by the three companies, the New York Times reported. The DOJ will lead the investigation …
…
continue reading

1
INTERVIEW: Scaling Democracy w/ (Dr.) Igor Krawczuk
2:58:46
2:58:46
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
2:58:46The almost Dr. Igor Krawczuk joins me for what is the equivalent of 4 of my previous episodes. We get into all the classics: eugenics, capitalism, philosophical toads... Need I say more? If you're interested in connecting with Igor, head on over to his website, or check out placeholder for thesis (it isn't published yet). Because the full show note…
…
continue reading

1
AISN #36: Voluntary Commitments are Insufficient
10:09
10:09
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
10:09Voluntary Commitments are Insufficient AI companies agree to RSPs in Seoul. Following the second AI Global Summit held in Seoul, the UK and Republic of Korea governments announced that 16 major technology organizations, including Amazon, Google, Meta, Microsoft, OpenAI, and xAI have agreed to a new set of Frontier AI Safety Commitments. Some commit…
…
continue reading
OpenAI and Google Announce New Multimodal Models In the current paradigm of AI development, there are long delays between the release of successive models. Progress is largely driven by increases in computing power, and training models with more computing power requires building large new data centers. More than a year after the release of GPT-4, O…
…
continue reading
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. AI Labs Fail to Uphold Safety Commitments to UK AI Safety Institute In November, leading AI labs committed to sharing their models before deployment to be tested by the UK AI Safety Institute. But reporting…
…
continue reading

1
AISN #33: Reassessing AI and Biorisk
20:27
20:27
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
20:27Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. This week, we cover: Consolidation in the corporate AI landscape, as smaller startups join forces with larger funders. Several countries have announced new investments in AI, including Singapore, Canada, an…
…
continue reading

1
INTERVIEW: StakeOut.AI w/ Dr. Peter Park (3)
1:42:00
1:42:00
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
1:42:00As always, the best things come in 3s: dimensions, musketeers, pyramids, and... 3 installments of my interview with Dr. Peter Park, an AI Existential Safety Post-doctoral Fellow working with Dr. Max Tegmark at MIT. As you may have ascertained from the previous two segments of the interview, Dr. Park cofounded StakeOut.AI along with Harry Luk and on…
…
continue reading

1
INTERVIEW: StakeOut.AI w/ Dr. Peter Park (2)
1:06:23
1:06:23
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
1:06:23Join me for round 2 with Dr. Peter Park, an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. Dr. Park was a cofounder of StakeOut.AI, a non-profit focused on making AI go well for humans, along with Harry Luk and one other individual, whose name has been removed due to requirements of her current position. In addition …
…
continue reading
UPDATE: Contrary to what I say in this episode, I won't be removing any episodes that are already published from the podcast RSS feed. After getting some advice and reflecting more on my own personal goals, I have decided to shift the direction of the podcast towards accessible content regarding "AI" instead of the show's original focus. I will sti…
…
continue reading

1
AISN #32: Measuring and Reducing Hazardous Knowledge in LLMs
17:56
17:56
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
17:56Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. Measuring and Reducing Hazardous Knowledge The recent White House Executive Order on Artificial Intelligence highlights risks of LLMs in facilitating the development of bioweapons, chemical weapons, and cyb…
…
continue reading

1
INTERVIEW: StakeOut.AI w/ Dr. Peter Park (1)
54:11
54:11
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
54:11Dr. Peter Park is an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. In conjunction with Harry Luk and one other cofounder, he founded StakeOut.AI, a non-profit focused on making AI go well for humans. 00:54 - Intro 03:15 - Dr. Park, x-risk, and AGI 08:55 - StakeOut.AI 12:05 - Governance scorecard 19:34 - Hollywood w…
…
continue reading
Take a trip with me through the paper Large Language Models, A Survey, published on February 9th of 2024. All figures and tables mentioned throughout the episode can be found on the Into AI Safety podcast website. 00:36 - Intro and authors 01:50 - My takes and paper structure 04:40 - Getting to LLMs 07:27 - Defining LLMs & emergence 12:12 - Overvie…
…
continue reading

1
AISN #31: A New AI Policy Bill in California
13:24
13:24
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
13:24Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. This week, we’ll discuss: A new proposed AI bill in California which requires frontier AI developers to adopt safety and security protocols, and clarifies that developers bear legal liability if their AI sy…
…
continue reading

1
FEEDBACK: Applying for Funding w/ Esben Kran
45:13
45:13
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
45:13Esben reviews an application that I would soon submit for Open Philanthropy's Career Transitition Funding opportunity. Although I didn't end up receiving the funding, I do think that this episode can be a valuable resource for both others and myself when applying for funding in the future. Head over to Apart Research's website to check out their wo…
…
continue reading
Before I begin with the paper-distillation based minisodes, I figured we would go over best practices for reading research papers. I go through the anatomy of typical papers, and some generally applicable advice. 00:56 - Anatomy of a paper 02:38 - Most common advice 05:24 - Reading sparsity and path 07:30 - Notes and motivation Links to all article…
…
continue reading
Join our hackathon group for the second episode in the Evals November 2023 Hackathon subseries. In this episode, we solidify our goals for the hackathon after some preliminary experimentation and ideation. Check out Stellaric's website, or follow them on Twitter. 01:53 - Meeting starts 05:05 - Pitch: extension of locked models 23:23 - Pitch: retroa…
…
continue reading
I provide my thoughts and recommendations regarding personal professional portfolios. 00:35 - Intro to portfolios 01:42 - Modern portfolios 02:27 - What to include 04:38 - Importance of visual 05:50 - The "About" page 06:25 - Tools 08:12 - Future of "Minisodes" Links to all articles/papers which are mentioned throughout the episode can be found bel…
…
continue reading

1
AISN #30: Investments in Compute and Military AI
11:25
11:25
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
11:25Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. Compute Investments Continue To Grow Pausing AI development has been proposed as a policy for ensuring safety. For example, an open letter last year from the Future of Life Institute called for a six-month …
…
continue reading

1
INTERVIEW: Polysemanticity w/ Dr. Darryl Wright
45:09
45:09
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
45:09Darryl and I discuss his background, how he became interested in machine learning, and a project we are currently working on investigating the penalization of polysemanticity during the training of neural networks. Check out a diagram of the decoder task used for our research! 01:46 - Interview begins 02:14 - Supernovae classification 08:58 - Penal…
…
continue reading
A summary and reflections on the path I have taken to get this podcast started, including some resources recommendations for others who want to do something similar. Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. LessWrong Spotify for Podcasters Into AI Safety podcast websit…
…
continue reading

1
HACKATHON: Evals November 2023 (1)
1:08:39
1:08:39
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
1:08:39This episode kicks off our first subseries, which will consist of recordings taken during my team's meetings for the AlignmentJams Evals Hackathon in November of 2023. Our team won first place, so you'll be listening to the process which, at the end of the day, turned out to be pretty good. Check out Apart Research, the group that runs the Alignmen…
…
continue reading
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. A Provisional Agreement on the EU AI Act On December 8th, the EU Parliament, Council, and Commission reached a provisional agreement on the EU AI Act. The agreement regulates the deployment of AI in high ri…
…
continue reading
In this minisode I give some tips for staying up-to-date in the everchanging landscape of AI. I would like to point out that I am constantly iterating on these strategies, tools, and sources, so it is likely that I will make an update episode in the future. Links to all articles/papers which are mentioned throughout the episode can be found below, …
…
continue reading
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. This week we’re looking closely at AI legislative efforts in the United States, including: Senator Schumer's AI Insight Forum The Blumenthal-Hawley framework for AI governance Agencies proposed to govern di…
…
continue reading

1
AISN #28: Center for AI Safety 2023 Year in Review
11:08
11:08
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
11:08As 2023 comes to a close, we want to thank you for your continued support for AI safety. This has been a big year for AI and for the Center for AI Safety. In this special-edition newsletter, we highlight some of our most important projects from the year. Thank you for being part of our community and our work. Center for AI Safety's 2023 Year in Rev…
…
continue reading