pod.link/1630783021
pod.link copied!
LessWrong (Curated & Popular)
LessWrong (Curated & Popular)
LessWrong

Audio narrations of LessWrong posts. Includes all curated posts and all posts with 125+ karma.If you'd like more, subscribe to the “Lesswrong... more

Listen now on

Apple Podcasts
Spotify
Overcast
Podcast Addict
Pocket Casts
Castbox
Podbean
iHeartRadio
Player FM
Podcast Republic
Castro
RSS

Episodes

“The Field of AI Alignment: A Postmortem, and What To Do About It” by johnswentworth

A policeman sees a drunk man searching for something under a streetlight and asks what the drunk has lost. He... more

26 Dec 2024 · 14 minutes
“When Is Insurance Worth It?” by kqr

TL;DR: If you want to know whether getting insurance is worth it, use the Kelly Insurance Calculator. If you want... more

23 Dec 2024 · 11 minutes
“Orienting to 3 year AGI timelines” by Nikola Jurkovic

My median expectation is that AGI[1] will be created 3 years from now. This has implications on how to behave,... more

23 Dec 2024 · 14 minutes
“What Goes Without Saying” by sarahconstantin

There are people I can talk to, where all of the following statements are obvious. They go without saying. We... more

21 Dec 2024 · 9 minutes
“o3” by Zach Stein-Perlman

I'm editing this post.OpenAI announced (but hasn't released) o3 (skipping o2 for trademark reasons).It gets 25% on FrontierMath, smashing the... more

21 Dec 2024 ·
“‘Alignment Faking’ frame is somewhat fake” by Jan_Kulveit

I like the research. I mostly trust the results. I dislike the 'Alignment Faking' name and frame, and I'm afraid... more

21 Dec 2024 · 11 minutes
“AIs Will Increasingly Attempt Shenanigans” by Zvi

Increasingly, we have seen papers eliciting in AI models various shenanigans.There are a wide variety of scheming behaviors. You’ve got... more

19 Dec 2024 · 51 minutes
“Alignment Faking in Large Language Models” by ryan_greenblatt, evhub, Carson Denison, Benjamin Wright, Fabien Roger, Monte M, Sam Marks, Johannes Treutlein, Sam Bowman, Buck

What happens when you tell Claude it is being trained to do something it doesn't want to do? We (Anthropic... more

18 Dec 2024 · 19 minutes
“Communications in Hard Mode (My new job at MIRI)” by tanagrabeast

Six months ago, I was a high school English teacher.I wasn’t looking to change careers, even after nineteen sometimes-difficult years.... more

15 Dec 2024 · 10 minutes
“Biological risk from the mirror world” by jasoncrawford

A new article in Science Policy Forum voices concern about a particular line of biological research which, if successful in... more

13 Dec 2024 · 14 minutes
LessWrong (Curated & Popular)
“The Field of AI Alignment: A Postmortem, and What To Do About It” by johnswentworth
LessWrong (Curated & Popular)
Claim your free pod.link
Customize to match your brand
Claim a memorable URL
Add your own Google Analytics
Confirm Ownership

To claim this podcast, you must confirm your ownership via the email address located in your podcast’s RSS feed (). If you cannot access this email, please contact your hosting provider.