pod.link copied!
AI Safety Fundamentals: Alignment
AI Safety Fundamentals: Alignment
BlueDot Impact

Listen to resources from the AI Safety Fundamentals: Alignment course!https://aisafetyfundamentals.com/alignment

Listen now on

Apple Podcasts
Google Podcasts
Podcast Addict
Pocket Casts
Player FM
Podcast Republic


How to get feedback

Feedback is essential for learning. Whether you’re studying for a test, trying to improve in your work or want to... more

12 May 2024 · 7 minutes
Public by default: How we manage information visibility at Get on Board

I’ve been obsessed with managing information, and communications in a remote team since Get on Board started growing. Reducing the... more

12 May 2024 · 9 minutes
Writing, Briefly

(In the process of answering an email, I accidentally wrote a tiny essay about writing. I usually spend weeks on... more

12 May 2024 · 3 minutes
Being the (Pareto) Best in the World

This introduces the concept of Pareto frontiers. The top comment by Rob Miles also ties it to comparative advantage.While reading,... more

04 May 2024 · 6 minutes
How to succeed as an early-stage researcher: the “lean startup” approach

I am approaching the end of my AI governance PhD, and I’ve spent about 2.5 years as a researcher at... more

23 Apr 2024 · 15 minutes
Become a person who Actually Does Things

The next four weeks of the course are an opportunity for you to actually build a thing that moves you... more

17 Apr 2024 · 5 minutes
Planning a High-Impact Career: A Summary of Everything You Need to Know in 7 Points

We took 10 years of research and what we’ve learned from advising 1,000+ people on how to build high-impact careers,... more

16 Apr 2024 · 11 minutes
Working in AI Alignment

This guide is written for people who are considering direct work on technical AI alignment. I expect it to be... more

14 Apr 2024 · 1 hour, 8 minutes
Computing Power and the Governance of AI

This post summarises a new report, “Computing Power and the Governance of Artificial Intelligence.” The full report is a collaboration... more

07 Apr 2024 · 26 minutes
AI Control: Improving Safety Despite Intentional Subversion

We’ve released a paper, AI Control: Improving Safety Despite Intentional Subversion. This paper explores techniques that prevent AI catastrophes even... more

07 Apr 2024 · 20 minutes
AI Safety Fundamentals: Alignment
How to get feedback
AI Safety Fundamentals: Alignment