Log in

goodpods headphones icon

To access all our features

Open the Goodpods app
Close icon
headphones
Dwarkesh Podcast

Dwarkesh Podcast

Dwarkesh Patel

Deeply researched interviews
www.dwarkeshpatel.com
profile image

3 Listeners

bookmark
Share icon

All episodes

Best episodes

Top 10 Dwarkesh Podcast Episodes

Goodpods has curated a list of the 10 best Dwarkesh Podcast episodes, ranked by the number of listens and likes each episode have garnered from our listeners. If you are listening to Dwarkesh Podcast for the first time, there's no better place to start than with one of these standout episodes. If you are a fan of the show, vote for your favorite Dwarkesh Podcast episode by adding your comments to the episode page.

In terms of the depth and range of topics, this episode is the best I’ve done.

No part of my worldview is the same after talking with Carl Shulman. He's the most interesting intellectual you've never heard of.

We ended up talking for 8 hours, so I'm splitting this episode into 2 parts.

This part is about Carl’s model of an intelligence explosion, which integrates everything from:

how fast algorithmic progress & hardware improvements in AI are happening,

what primate evolution suggests about the scaling hypothesis,

how soon before AIs could do large parts of AI research themselves, and whether there would be faster and faster doublings of AI researchers,

how quickly robots produced from existing factories could take over the economy.

We also discuss the odds of a takeover based on whether the AI is aligned before the intelligence explosion happens, and Carl explains why he’s more optimistic than Eliezer.

The next part, which I’ll release next week, is about all the specific mechanisms of an AI takeover, plus a whole bunch of other galaxy brain stuff.

Maybe 3 people in the world have thought as rigorously as Carl about so many interesting topics. This was a huge pleasure.

Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.

Timestamps

(00:00:00) - Intro

(00:01:32) - Intelligence Explosion

(00:18:03) - Can AIs do AI research?

(00:39:00) - Primate evolution

(01:03:30) - Forecasting AI progress

(01:34:20) - After human-level AGI

(02:08:39) - AI takeover scenarios


Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
profile image

2 Listeners

bookmark
plus icon
share episode

I went over to the OpenAI offices in San Fransisco to ask the Chief Scientist and cofounder of OpenAI, Ilya Sutskever, about:

time to AGI

leaks and spies

what's after generative models

post AGI futures

working with Microsoft and competing with Google

difficulty of aligning superhuman AI

Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.

Timestamps

(00:00) - Time to AGI

(05:57) - What’s after generative models?

(10:57) - Data, models, and research

(15:27) - Alignment

(20:53) - Post AGI Future

(26:56) - New ideas are overrated

(36:22) - Is progress inevitable?

(41:27) - Future Breakthroughs


Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

2 Listeners

bookmark
plus icon
share episode

Here is my episode with Demis Hassabis, CEO of Google DeepMind

We discuss:

Why scaling is an artform

Adding search, planning, & AlphaZero type training atop LLMs

Making sure rogue nations can't steal weights

The right way to align superhuman AIs and do an intelligence explosion

Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here.

Timestamps

(0:00:00) - Nature of intelligence

(0:05:56) - RL atop LLMs

(0:16:31) - Scaling and alignment

(0:24:13) - Timelines and intelligence explosion

(0:28:42) - Gemini training

(0:35:30) - Governance of superhuman AIs

(0:40:42) - Safety, open source, and security of weights

(0:47:00) - Multimodal and further progress

(0:54:18) - Inside Google DeepMind


Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

1 Listener

bookmark
plus icon
share episode

For 4 hours, I tried to come up reasons for why AI might not kill us all, and Eliezer Yudkowsky explained why I was wrong.

We also discuss his call to halt AI, why LLMs make alignment harder, what it would take to save humanity, his millions of words of sci-fi, and much more.

If you want to get to the crux of the conversation, fast forward to 2:35:00 through 3:43:54. Here we go through and debate the main reasons I still think doom is unlikely.

Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.

Timestamps

(0:00:00) - TIME article

(0:09:06) - Are humans aligned?

(0:37:35) - Large language models

(1:07:15) - Can AIs help with alignment?

(1:30:17) - Society’s response to AI

(1:44:42) - Predictions (or lack thereof)

(1:56:55) - Being Eliezer

(2:13:06) - Othogonality

(2:35:00) - Could alignment be easier than we think?

(3:02:15) - What will AIs want?

(3:43:54) - Writing fiction & whether rationality helps you win


Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

1 Listener

bookmark
plus icon
share episode

Here is my conversation with Dario Amodei, CEO of Anthropic.

Dario is hilarious and has fascinating takes on what these models are doing, why they scale so well, and what it will take to align them.

Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.

Timestamps

(00:00:00) - Introduction

(00:01:00) - Scaling

(00:15:46) - Language

(00:22:58) - Economic Usefulness

(00:38:05) - Bioterrorism

(00:43:35) - Cybersecurity

(00:47:19) - Alignment & mechanistic interpretability

(00:57:43) - Does alignment research require scale?

(01:05:30) - Misuse vs misalignment

(01:09:06) - What if AI goes well?

(01:11:05) - China

(01:15:11) - How to think about alignment

(01:31:31) - Is modern security good enough?

(01:36:09) - Inefficiencies in training

(01:45:53) - Anthropic’s Long Term Benefit Trust

(01:51:18) - Is Claude conscious?

(01:56:14) - Keeping a low profile


Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

1 Listener

bookmark
plus icon
share episode

I learned so much from Sarah Paine, Professor of History and Strategy at the Naval War College.

We discuss:

how continental vs maritime powers think and how this explains Xi & Putin's decisions

how a war with China over Taiwan would shake out and whether it could go nuclear

why the British Empire fell apart, why China went communist, how Hitler and Japan could have coordinated to win WW2, and whether Japanese occupation was good for Korea, Taiwan and Manchuria

plus other lessons from WW2, Cold War, and Sino-Japanese War

how to study history properly, and why leaders keep making the same mistakes

If you want to learn more, check out her books - they’re some of the best military history I’ve ever read.

Watch on YouTube, listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript.

Timestamps

(0:00:00) - Grand strategy

(0:11:59) - Death ground

(0:23:19) - WW1

(0:39:23) - Writing history

(0:50:25) - Japan in WW2

(0:59:58) - Ukraine

(1:10:50) - Japan/Germany vs Iraq/Afghanistan occupation

(1:21:25) - Chinese invasion of Taiwan

(1:51:26) - Communists & Axis

(2:08:34) - Continental vs maritime powers


Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

1 Listener

bookmark
plus icon
share episode
Will MacAskill is one of the founders of the Effective Altruist movement and the author of the upcoming book, What We Owe The Future.

We talk about improving the future, risk of extinction & collapse, technological & moral change, problems of academia, who changes history, and much more.

Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform.

Read the full transcript here.

Follow Will on Twitter. Follow me on Twitter for updates on future episodes.

Timestamps

(00:23) - Effective Altruism and Western values

(07:47) - The contingency of technology

(12:02) - Who changes history?

(18:00) - Longtermist institutional reform

(25:56) - Are companies longtermist?

(28:57) - Living in an era of plasticity

(34:52) - How good can the future be?

(39:18) - Contra Tyler Cowen on what’s most important

(45:36) - AI and the centralization of power

(51:34) - The problems with academia

This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit www.dwarkeshpatel.com
profile image

1 Listener

bookmark
plus icon
share episode
Dwarkesh Podcast - Bryan Caplan - Nurturing Orphaned Ideas
play

05/22/20 • 59 min

Bryan Caplan is a Professor of Economics at George Mason University and a New York Times Bestselling author. His most famous works include: The Myth of the Rational Voter, Selfish Reasons to Have More Kids, The Case Against Education, and Open Borders: The Science and Ethics of Immigration.
I talk to Bryan about open borders, the idea trap, UBI, appeasement, China, the education system, and Bryan Caplan's next two books on poverty and housing regulation.

Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform.
Follow Bryan on Twitter. Follow me on Twitter for updates on future episodes.


Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
profile image

1 Listener

bookmark
plus icon
share episode

It was a tremendous honor & pleasure to interview Richard Rhodes, Pulitzer Prize winning author of The Making of the Atomic Bomb

We discuss

similarities between AI progress & Manhattan Project (developing a powerful, unprecedented, & potentially apocalyptic technology within an uncertain arms-race situation)

visiting starving former Soviet scientists during fall of Soviet Union

whether Oppenheimer was a spy, & consulting on the Nolan movie

living through WW2 as a child

odds of nuclear war in Ukraine, Taiwan, Pakistan, & North Korea

how the US pulled of such a massive secret wartime scientific & industrial project

Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.

Timestamps

(0:00:00) - Oppenheimer movie

(0:06:22) - Was the bomb inevitable?

(0:29:10) - Firebombing vs nuclear vs hydrogen bombs

(0:49:44) - Stalin & the Soviet program

(1:08:24) - Deterrence, disarmament, North Korea, Taiwan

(1:33:12) - Oppenheimer as lab director

(1:53:40) - AI progress vs Manhattan Project

(1:59:50) - Living through WW2

(2:16:45) - Secrecy

(2:26:34) - Wisdom & war


Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

1 Listener

bookmark
plus icon
share episode
Dwarkesh Podcast - Paul Christiano - Preventing an AI Takeover
play

10/31/23 • 187 min

Paul Christiano is the world’s leading AI safety researcher. My full episode with him is out!

We discuss:

Does he regret inventing RLHF, and is alignment necessarily dual-use?

Why he has relatively modest timelines (40% by 2040, 15% by 2030),

What do we want post-AGI world to look like (do we want to keep gods enslaved forever)?

Why he’s leading the push to get to labs develop responsible scaling policies, and what it would take to prevent an AI coup or bioweapon,

His current research into a new proof system, and how this could solve alignment by explaining model's behavior

and much more.

Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes.

Open Philanthropy

Open Philanthropy is currently hiring for twenty-two different roles to reduce catastrophic risks from fast-moving advances in AI and biotechnology, including grantmaking, research, and operations.

For more information and to apply, please see the application: https://www.openphilanthropy.org/research/new-roles-on-our-gcr-team/

The deadline to apply is November 9th; make sure to check out those roles before they close.

Timestamps

(00:00:00) - What do we want post-AGI world to look like?

(00:24:25) - Timelines

(00:45:28) - Evolution vs gradient descent

(00:54:53) - Misalignment and takeover

(01:17:23) - Is alignment dual-use?

(01:31:38) - Responsible scaling policies

(01:58:25) - Paul’s alignment research

(02:35:01) - Will this revolutionize theoretical CS and math?

(02:46:11) - How Paul invented RLHF

(02:55:10) - Disagreements with Carl Shulman

(03:01:53) - Long TSMC but not NVIDIA


Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe

1 Listener

bookmark
plus icon
share episode

Show more best episodes

Toggle view more icon

Featured in these lists

FAQ

How many episodes does Dwarkesh Podcast have?

Dwarkesh Podcast currently has 90 episodes available.

What topics does Dwarkesh Podcast cover?

The podcast is about Podcasts, Technology and Science.

What is the most popular episode on Dwarkesh Podcast?

The episode title 'Ilya Sutskever (OpenAI Chief Scientist) - Building AGI, Alignment, Future Models, Spies, Microsoft, Taiwan, & Enlightenment' is the most popular.

What is the average episode length on Dwarkesh Podcast?

The average episode length on Dwarkesh Podcast is 96 minutes.

How often are episodes of Dwarkesh Podcast released?

Episodes of Dwarkesh Podcast are typically released every 10 days.

When was the first episode of Dwarkesh Podcast?

The first episode of Dwarkesh Podcast was released on May 22, 2020.

Show more FAQ

Toggle view more icon

Comments