pod.link/1565088425
pod.link copied!
The Inside View
The Inside View
Michaël Trazzi

The goal of this podcast is to create a place where people discuss their inside views about existential risk from AI.

Listen now on

Apple Podcasts
Spotify
Google Podcasts
Overcast
Podcast Addict
Pocket Casts
Castbox
Stitcher
Podbean
iHeartRadio
Player FM
Podcast Republic
Castro
RadioPublic
RSS

Episodes

Ethan Perez on Selecting Alignment Research Projects (ft. Mikita Balesni & Henry Sleight)

Ethan Perez is a Research Scientist at Anthropic, where he leads a team working on developing model organisms of misalignment. Youtube:... more

09 Apr 2024 · 36 minutes
Emil Wallner on Sora, Generative AI Startups and AI optimism

Emil is the co-founder of palette.fm (colorizing B&W pictures with generative AI) and was previously working in deep learning for... more

20 Feb 2024 · 1 hour, 42 minutes
Evan Hubinger on Sleeper Agents, Deception and Responsible Scaling Policies

Evan Hubinger leads the Alignment stress-testing at Anthropic and recently published "Sleeper Agents: Training Deceptive LLMs That Persist Through Safety... more

12 Feb 2024 · 52 minutes
[Jan 2023] Jeffrey Ladish on AI Augmented Cyberwarfare and compute monitoring

Jeffrey Ladish is the Executive Director of Palisade Research which aimes so "study the offensive capabilities or AI systems today... more

27 Jan 2024 · 33 minutes
Holly Elmore on pausing AI

Holly Elmore is an AI Pause Advocate who has organized two protests in the past few months (against Meta's open... more

22 Jan 2024 · 1 hour, 40 minutes
Podcast Retrospective and Next Steps

https://youtu.be/Fk2MrpuWinc

09 Jan 2024 · 1 hour, 3 minutes
Paul Christiano's views on "doom" (ft. Robert Miles)

Youtube: https://youtu.be/JXYcLQItZsk Paul Christiano's post: https://www.lesswrong.com/posts/xWMqsvHapP3nwdSW8/my-views-on-doom

29 Sep 2023 · 4 minutes
Neel Nanda on mechanistic interpretability, superposition and grokking

Neel Nanda is a researcher at Google DeepMind working on mechanistic interpretability. He is also known for his YouTube channel... more

21 Sep 2023 · 2 hours, 4 minutes
Joscha Bach on how to stop worrying and love AI

Joscha Bach (who defines himself as an AI researcher/cognitive scientist) has recently been debating existential risk from AI with Connor... more

08 Sep 2023 · 2 hours, 54 minutes
Erik Jones on Automatically Auditing Large Language Models

Erik is a Phd at Berkeley working with Jacob Steinhardt, interested in making generative machine learning systems more robust, reliable,... more

11 Aug 2023 · 22 minutes
Dylan Patel on the GPU Shortage, Nvidia and the Deep Learning Supply Chain

Dylan Patel is Chief Analyst at SemiAnalysis a boutique semiconductor research and consulting firm specializing in the semiconductor supply chain from... more

09 Aug 2023 · 12 minutes
Tony Wang on Beating Superhuman Go AIs with Advesarial Policies

Tony is a PhD student at MIT, and author of "Advesarial Policies Beat Superhuman Go AIs", accepted as Oral at... more

04 Aug 2023 · 3 minutes
David Bau on Editing Facts in GPT, AI Safety and Interpretability

David Bau is an Assistant Professor studying the structure and interpretation of deep networks, and the co-author on "Locating and... more

01 Aug 2023 · 24 minutes
Alexander Pan on the MACHIAVELLI benchmark

I've talked to Alexander Pan, 1st year at Berkeley working with Jacob Steinhardt about his paper "Measuring Trade-Offs Between Rewards... more

26 Jul 2023 · 20 minutes
Vincent Weisser on Funding AI Alignment Research

Vincent is currently spending his time supporting AI alignment efforts, as well as investing across AI, semi, energy, crypto, bio... more

24 Jul 2023 · 18 minutes
[JUNE 2022] Aran Komatsuzaki on Scaling, GPT-J and Alignment

Aran Komatsuzaki is a ML PhD student at GaTech and lead researcher at EleutherAI where he was one of the... more

19 Jul 2023 · 1 hour, 17 minutes
Nina Rimsky on AI Deception and Mesa-optimisation

Nina is a software engineer at Stripe currently working with Evan Hubinger (Anthropic) on AI Deception and Mesa Optimization. I... more

18 Jul 2023 · 55 minutes
Curtis Huebner on Doom, AI Timelines and Alignment at EleutherAI

Curtis, also known on the internet as AI_WAIFU, is the head of Alignment at EleutherAI. In this episode we discuss... more

16 Jul 2023 · 1 hour, 29 minutes
Eric Michaud on scaling, grokking and quantum interpretability

Eric is a PhD student in the Department of Physics at MIT working with Max Tegmark on improving our scientific/theoretical... more

12 Jul 2023 · 48 minutes
Jesse Hoogland on Developmental Interpretability and Singular Learning Theory

Jesse Hoogland is a research assistant at David Krueger's lab in Cambridge studying AI Safety. More recently, Jesse has been... more

06 Jul 2023 · 43 minutes
Clarifying and predicting AGI by Richard Ngo

Explainer podcast for Richard Ngo's "Clarifying and predicting AGI" post on Lesswrong, which introduces the t-AGI framework to evaluate AI... more

09 May 2023 · 4 minutes
Alan Chan And Max Kauffman on Model Evaluations, Coordination and AI Safety

Max Kaufmann⁠ and Alan Chan discuss the evaluation of large language models, AI Governance and more generally the impact of... more

06 May 2023 · 1 hour, 13 minutes
Breandan Considine on Neuro Symbolic AI, Coding AIs and AI Timelines

Breandan Considine is a PhD student at the School of Computer Science at McGill University, under the supervision of Jin... more

04 May 2023 · 1 hour, 45 minutes
Christoph Schuhmann on Open Source AI, Misuse and Existential risk

Christoph Schuhmann is the co-founder and organizational lead at LAION, the non-profit who released LAION-5B, a dataset of 5,85 billion... more

01 May 2023 · 32 minutes
Simeon Campos on Short Timelines, AI Governance and AI Alignment Field Building

Siméon Campos is the founder of EffiSciences and SaferAI, mostly focusing on alignment field building and AI Governance. More recently,... more

29 Apr 2023 · 2 hours, 3 minutes
Collin Burns On Discovering Latent Knowledge In Language Models Without Supervision

Collin Burns is a second-year ML PhD at Berkeley, working with Jacob Steinhardt on making language models honest, interpretable, and... more

17 Jan 2023 · 2 hours, 34 minutes
Victoria Krakovna–AGI Ruin, Sharp Left Turn, Paradigms of AI Alignment

Victoria Krakovna is a Research Scientist at DeepMind working on AGI safety and a co-founder of the Future of Life... more

12 Jan 2023 · 1 hour, 52 minutes
David Krueger–Coordination, Alignment, Academia

David Krueger is an assistant professor at the University of Cambridge and got his PhD from Mila. His research group... more

07 Jan 2023 · 2 hours, 45 minutes
Ethan Caballero–Broken Neural Scaling Laws

Ethan Caballero is a PhD student at Mila interested in how to best scale Deep Learning models according to all... more

03 Nov 2022 · 23 minutes
Irina Rish–AGI, Scaling and Alignment

Irina Rish a professor at the Université de Montréal, a core member of Mila (Quebec AI Institute), and the organizer... more

18 Oct 2022 · 1 hour, 26 minutes
The Inside View
Victoria Krakovna–AGI Ruin, Sharp Left Turn, Paradigms of AI Alignment
The Inside View
0:00
-0:00

Listen now on

Apple Podcasts
Spotify
Google Podcasts
Overcast
Podcast Addict
Pocket Casts
Castbox
Stitcher
Podbean
iHeartRadio
Player FM
Podcast Republic
Castro
RadioPublic
RSS

Description

Victoria Krakovna is a Research Scientist at DeepMind working on AGI safety and a co-founder of the Future of Life Institute, a... more