Ilya & I discuss SSIβs strategy, the problems with pre-training, how to improve the generalization of AI models, and how to ensure AGI goes well.
πππππππ πππππ
* Transcript: https://www.dwarkesh.com/p/ilya-sutskever-2
* Apple Podcasts: https://podcasts.apple.com/us/podcast/dwarkesh-podcast/id1516093381?i=1000738363711
* Spotify: https://open.spotify.com/episode/7naOOba8SwiUNobGz8mQEL?si=39dd68f346ea4d49
ππππππππ
– Gemini 3 is the first model Iβve used that can find connections I havenβt anticipated. I recently wrote a blog post on RLβs information efficiency, and Gemini 3 helped me think it all through. It also generated the relevant charts and ran toy ML experiments for me with zero bugs. Try Gemini 3 today at https://gemini.google
– Labelbox helped me create a tool to transcribe our episodes! Iβve struggled with transcription in the past because I donβt just want verbatim transcripts, I want transcripts reworded to read like essays. Labelbox helped me generate the *exact* data I needed for this. If you want to learn how Labelbox can help you (or if you want to try out the transcriber tool yourself), go to https://labelbox.com/dwarkesh
– Sardine is an AI risk management platform that brings together thousands of device, behavior, and identity signals to help you assess a userβs risk of fraud & abuse. Sardine also offers a suite of agents to automate investigations so that as fraudsters use AI to scale their attacks, you can use AI to scale your defenses. Learn more at https://sardine.ai/dwarkesh
To sponsor a future episode, visit https://dwarkesh.com/advertise
ππππππππππ
00:00:00 β Explaining model jaggedness
00:09:39 – Emotions and value functions
00:18:49 β What are we scaling?
00:25:13 β Why humans generalize better than models
00:35:45 β Straight-shotting superintelligence
00:46:47 β SSIβs model will learn from deployment
00:55:07 β Alignment
01:18:13 β βWe are squarely an age of research companyβ
01:29:23 — Self-play and multi-agent
01:32:42 β Research taste
source
