Search

Home > The Bayesian Conspiracy > 213 – Are Transformer Models Aligned By Default?
Podcast: The Bayesian Conspiracy
Episode:

213 – Are Transformer Models Aligned By Default?

Category: Education
Duration: 00:00:00
Publish Date: 2024-05-29 08:01:51
Description:

Our species has begun to scrute the inscrutable shoggoth! With Matt Freeman

LINKS
Anthropic’s latest AI Safety research paper, on interpretability
Anthropic is hiring
Episode 93 of The Mind Killer
Talkin’ Fallout
VibeCamp

0:00:17 – A Layman’s AI Refresher
0:21:06 – Aligned By Default
0:50:56 – Highlights from Anthropic’s Latest Interpretability Paper
1:26:47 – Guild of the Rose Update
1:29:40 – Going to VibeCamp
1:37:05 – Feedback
1:43:58 – Less Wrong Posts
1:57:30 – Thank the Patron


Our Patreon, or if you prefer Our SubStack

Hey look, we have a discord! What could possibly go wrong?
We now partner with The Guild of the Rose, check them out.

Total Play: 0