Dwarkesh Podcast
Dwarkesh Patel
www.dwarkeshpatel.com
3 Listeners
All episodes
Best episodes
Top 10 Dwarkesh Podcast Episodes
Goodpods has curated a list of the 10 best Dwarkesh Podcast episodes, ranked by the number of listens and likes each episode have garnered from our listeners. If you are listening to Dwarkesh Podcast for the first time, there's no better place to start than with one of these standout episodes. If you are a fan of the show, vote for your favorite Dwarkesh Podcast episode by adding your comments to the episode page.
Carl Shulman (Pt 1) - Intelligence Explosion, Primate Evolution, Robot Doublings, & Alignment
Dwarkesh Podcast
06/14/23 • 164 min
In terms of the depth and range of topics, this episode is the best I’ve done.
No part of my worldview is the same after talking with Carl Shulman. He's the most interesting intellectual you've never heard of.
We ended up talking for 8 hours, so I'm splitting this episode into 2 parts.
This part is about Carl’s model of an intelligence explosion, which integrates everything from:
how fast algorithmic progress & hardware improvements in AI are happening,
what primate evolution suggests about the scaling hypothesis,
how soon before AIs could do large parts of AI research themselves, and whether there would be faster and faster doublings of AI researchers,
how quickly robots produced from existing factories could take over the economy.
We also discuss the odds of a takeover based on whether the AI is aligned before the intelligence explosion happens, and Carl explains why he’s more optimistic than Eliezer.
The next part, which I’ll release next week, is about all the specific mechanisms of an AI takeover, plus a whole bunch of other galaxy brain stuff.
Maybe 3 people in the world have thought as rigorously as Carl about so many interesting topics. This was a huge pleasure.
Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.
Timestamps
(00:00:00) - Intro
(00:01:32) - Intelligence Explosion
(00:18:03) - Can AIs do AI research?
(00:39:00) - Primate evolution
(01:03:30) - Forecasting AI progress
(01:34:20) - After human-level AGI
(02:08:39) - AI takeover scenarios
Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
2 Listeners
Ilya Sutskever (OpenAI Chief Scientist) - Building AGI, Alignment, Future Models, Spies, Microsoft, Taiwan, & Enlightenment
Dwarkesh Podcast
03/27/23 • 47 min
I went over to the OpenAI offices in San Fransisco to ask the Chief Scientist and cofounder of OpenAI, Ilya Sutskever, about:
time to AGI
leaks and spies
what's after generative models
post AGI futures
working with Microsoft and competing with Google
difficulty of aligning superhuman AI
Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.
Timestamps
(00:00) - Time to AGI
(05:57) - What’s after generative models?
(10:57) - Data, models, and research
(15:27) - Alignment
(20:53) - Post AGI Future
(26:56) - New ideas are overrated
(36:22) - Is progress inevitable?
(41:27) - Future Breakthroughs
Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
2 Listeners
02/28/24 • 61 min
Here is my episode with Demis Hassabis, CEO of Google DeepMind
We discuss:
Why scaling is an artform
Adding search, planning, & AlphaZero type training atop LLMs
Making sure rogue nations can't steal weights
The right way to align superhuman AIs and do an intelligence explosion
Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here.
Timestamps
(0:00:00) - Nature of intelligence
(0:05:56) - RL atop LLMs
(0:16:31) - Scaling and alignment
(0:24:13) - Timelines and intelligence explosion
(0:28:42) - Gemini training
(0:35:30) - Governance of superhuman AIs
(0:40:42) - Safety, open source, and security of weights
(0:47:00) - Multimodal and further progress
(0:54:18) - Inside Google DeepMind
Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
1 Listener
Eliezer Yudkowsky - Why AI Will Kill Us, Aligning LLMs, Nature of Intelligence, SciFi, & Rationality
Dwarkesh Podcast
04/06/23 • 243 min
For 4 hours, I tried to come up reasons for why AI might not kill us all, and Eliezer Yudkowsky explained why I was wrong.
We also discuss his call to halt AI, why LLMs make alignment harder, what it would take to save humanity, his millions of words of sci-fi, and much more.
If you want to get to the crux of the conversation, fast forward to 2:35:00 through 3:43:54. Here we go through and debate the main reasons I still think doom is unlikely.
Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.
Timestamps
(0:00:00) - TIME article
(0:09:06) - Are humans aligned?
(0:37:35) - Large language models
(1:07:15) - Can AIs help with alignment?
(1:30:17) - Society’s response to AI
(1:44:42) - Predictions (or lack thereof)
(1:56:55) - Being Eliezer
(2:13:06) - Othogonality
(2:35:00) - Could alignment be easier than we think?
(3:02:15) - What will AIs want?
(3:43:54) - Writing fiction & whether rationality helps you win
Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
1 Listener
Dario Amodei (Anthropic CEO) - Scaling, Alignment, & AI Progress
Dwarkesh Podcast
08/08/23 • 118 min
Here is my conversation with Dario Amodei, CEO of Anthropic.
Dario is hilarious and has fascinating takes on what these models are doing, why they scale so well, and what it will take to align them.
Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.
Timestamps
(00:00:00) - Introduction
(00:01:00) - Scaling
(00:15:46) - Language
(00:22:58) - Economic Usefulness
(00:38:05) - Bioterrorism
(00:43:35) - Cybersecurity
(00:47:19) - Alignment & mechanistic interpretability
(00:57:43) - Does alignment research require scale?
(01:05:30) - Misuse vs misalignment
(01:09:06) - What if AI goes well?
(01:11:05) - China
(01:15:11) - How to think about alignment
(01:31:31) - Is modern security good enough?
(01:36:09) - Inefficiencies in training
(01:45:53) - Anthropic’s Long Term Benefit Trust
(01:51:18) - Is Claude conscious?
(01:56:14) - Keeping a low profile
Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
1 Listener
10/04/23 • 144 min
I learned so much from Sarah Paine, Professor of History and Strategy at the Naval War College.
We discuss:
how continental vs maritime powers think and how this explains Xi & Putin's decisions
how a war with China over Taiwan would shake out and whether it could go nuclear
why the British Empire fell apart, why China went communist, how Hitler and Japan could have coordinated to win WW2, and whether Japanese occupation was good for Korea, Taiwan and Manchuria
plus other lessons from WW2, Cold War, and Sino-Japanese War
how to study history properly, and why leaders keep making the same mistakes
If you want to learn more, check out her books - they’re some of the best military history I’ve ever read.
Watch on YouTube, listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript.
Timestamps
(0:00:00) - Grand strategy
(0:11:59) - Death ground
(0:23:19) - WW1
(0:39:23) - Writing history
(0:50:25) - Japan in WW2
(0:59:58) - Ukraine
(1:10:50) - Japan/Germany vs Iraq/Afghanistan occupation
(1:21:25) - Chinese invasion of Taiwan
(1:51:26) - Communists & Axis
(2:08:34) - Continental vs maritime powers
Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
1 Listener
07/26/22 • 56 min
We talk about improving the future, risk of extinction & collapse, technological & moral change, problems of academia, who changes history, and much more.
Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform.
Read the full transcript here.
Follow Will on Twitter. Follow me on Twitter for updates on future episodes.
Timestamps
(00:23) - Effective Altruism and Western values
(07:47) - The contingency of technology
(12:02) - Who changes history?
(18:00) - Longtermist institutional reform
(25:56) - Are companies longtermist?
(28:57) - Living in an era of plasticity
(34:52) - How good can the future be?
(39:18) - Contra Tyler Cowen on what’s most important
(45:36) - AI and the centralization of power
(51:34) - The problems with academia
This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit www.dwarkeshpatel.com
1 Listener
Bryan Caplan - Nurturing Orphaned Ideas
Dwarkesh Podcast
05/22/20 • 59 min
Bryan Caplan is a Professor of Economics at George Mason University and a New York Times Bestselling author. His most famous works include: The Myth of the Rational Voter, Selfish Reasons to Have More Kids, The Case Against Education, and Open Borders: The Science and Ethics of Immigration.
I talk to Bryan about open borders, the idea trap, UBI, appeasement, China, the education system, and Bryan Caplan's next two books on poverty and housing regulation.
Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform.
Follow Bryan on Twitter. Follow me on Twitter for updates on future episodes.
Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
1 Listener
05/23/23 • 157 min
It was a tremendous honor & pleasure to interview Richard Rhodes, Pulitzer Prize winning author of The Making of the Atomic Bomb
We discuss
similarities between AI progress & Manhattan Project (developing a powerful, unprecedented, & potentially apocalyptic technology within an uncertain arms-race situation)
visiting starving former Soviet scientists during fall of Soviet Union
whether Oppenheimer was a spy, & consulting on the Nolan movie
living through WW2 as a child
odds of nuclear war in Ukraine, Taiwan, Pakistan, & North Korea
how the US pulled of such a massive secret wartime scientific & industrial project
Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.
Timestamps
(0:00:00) - Oppenheimer movie
(0:06:22) - Was the bomb inevitable?
(0:29:10) - Firebombing vs nuclear vs hydrogen bombs
(0:49:44) - Stalin & the Soviet program
(1:08:24) - Deterrence, disarmament, North Korea, Taiwan
(1:33:12) - Oppenheimer as lab director
(1:53:40) - AI progress vs Manhattan Project
(1:59:50) - Living through WW2
(2:16:45) - Secrecy
(2:26:34) - Wisdom & war
Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
1 Listener
Paul Christiano - Preventing an AI Takeover
Dwarkesh Podcast
10/31/23 • 187 min
Paul Christiano is the world’s leading AI safety researcher. My full episode with him is out!
We discuss:
Does he regret inventing RLHF, and is alignment necessarily dual-use?
Why he has relatively modest timelines (40% by 2040, 15% by 2030),
What do we want post-AGI world to look like (do we want to keep gods enslaved forever)?
Why he’s leading the push to get to labs develop responsible scaling policies, and what it would take to prevent an AI coup or bioweapon,
His current research into a new proof system, and how this could solve alignment by explaining model's behavior
and much more.
Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.
Open Philanthropy
Open Philanthropy is currently hiring for twenty-two different roles to reduce catastrophic risks from fast-moving advances in AI and biotechnology, including grantmaking, research, and operations.
For more information and to apply, please see the application: https://www.openphilanthropy.org/research/new-roles-on-our-gcr-team/
The deadline to apply is November 9th; make sure to check out those roles before they close.
Timestamps
(00:00:00) - What do we want post-AGI world to look like?
(00:24:25) - Timelines
(00:45:28) - Evolution vs gradient descent
(00:54:53) - Misalignment and takeover
(01:17:23) - Is alignment dual-use?
(01:31:38) - Responsible scaling policies
(01:58:25) - Paul’s alignment research
(02:35:01) - Will this revolutionize theoretical CS and math?
(02:46:11) - How Paul invented RLHF
(02:55:10) - Disagreements with Carl Shulman
(03:01:53) - Long TSMC but not NVIDIA
Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
1 Listener
Show more best episodes
Show more best episodes
Featured in these lists
FAQ
How many episodes does Dwarkesh Podcast have?
Dwarkesh Podcast currently has 90 episodes available.
What topics does Dwarkesh Podcast cover?
The podcast is about Podcasts, Technology and Science.
What is the most popular episode on Dwarkesh Podcast?
The episode title 'Ilya Sutskever (OpenAI Chief Scientist) - Building AGI, Alignment, Future Models, Spies, Microsoft, Taiwan, & Enlightenment' is the most popular.
What is the average episode length on Dwarkesh Podcast?
The average episode length on Dwarkesh Podcast is 96 minutes.
How often are episodes of Dwarkesh Podcast released?
Episodes of Dwarkesh Podcast are typically released every 10 days.
When was the first episode of Dwarkesh Podcast?
The first episode of Dwarkesh Podcast was released on May 22, 2020.
Show more FAQ
Show more FAQ