TalkRL: The Reinforcement Learning Podcast
Robin Ranjit Singh Chauhan
1 Listener
All episodes
Best episodes
Top 10 TalkRL: The Reinforcement Learning Podcast Episodes
Goodpods has curated a list of the 10 best TalkRL: The Reinforcement Learning Podcast episodes, ranked by the number of listens and likes each episode have garnered from our listeners. If you are listening to TalkRL: The Reinforcement Learning Podcast for the first time, there's no better place to start than with one of these standout episodes. If you are a fan of the show, vote for your favorite TalkRL: The Reinforcement Learning Podcast episode by adding your comments to the episode page.
Natasha Jaques
TalkRL: The Reinforcement Learning Podcast
08/09/19 • 50 min
Natasha Jaques is a PhD candidate at MIT working on affective and social intelligence. She has interned with DeepMind and Google Brain, and was an OpenAI Scholars mentor. Her paper “Social Influence as Intrinsic Motivation for Multi-Agent Deep Reinforcement Learning” received an honourable mention for best paper at ICML 2019.
Featured References
Social Influence as Intrinsic Motivation for Multi-Agent Deep Reinforcement LearningNatasha Jaques, Angeliki Lazaridou, Edward Hughes, Caglar Gulcehre, Pedro A. Ortega, DJ Strouse, Joel Z. Leibo, Nando de Freitas
Tackling climate change with Machine LearningDavid Rolnick, Priya L. Donti, Lynn H. Kaack, Kelly Kochanski, Alexandre Lacoste, Kris Sankaran, Andrew Slavin Ross, Nikola Milojevic-Dupont, Natasha Jaques, Anna Waldman-Brown, Alexandra Luccioni, Tegan Maharaj, Evan D. Sherwin, S. Karthik Mukkavilli, Konrad P. Kording, Carla Gomes, Andrew Y. Ng, Demis Hassabis, John C. Platt, Felix Creutzig, Jennifer Chayes, Yoshua Bengio
Additional References
- MIT Media Lab Flight Offsets, Caroline Jaffe, Juliana Cherston, Natasha Jaques
- Modeling Others using Oneself in Multi-Agent Reinforcement Learning,
Roberta Raileanu, Emily Denton, Arthur Szlam, Rob Fergus - Inequity aversion improves cooperation in intertemporal social dilemmas,
Edward Hughes, Joel Z. Leibo, Matthew G. Phillips, Karl Tuyls, Edgar A. Duéñez-Guzmán, Antonio García Castañeda, Iain Dunning, Tina Zhu, Kevin R. McKee, Raphael Koster, Heather Roff, Thore Graepel - Sequential Social Dilemma Games on github, Eugene Vinitsky, Natasha Jaques
- AI Alignment newsletter, Rohin Shah
- Paired Open-Ended Trailblazer (POET): Endlessly Generating Increasingly Complex and Diverse Learning Environments and Their Solutions, Rui Wang, Joel Lehman, Jeff Clune, Kenneth O. Stanley
- The social function of intellect, Nicholas Humphrey
- Autocurricula and the Emergence of Innovation from Social Interaction: A Manifesto for Multi-Agent Intelligence Research, Joel Z. Leibo, Edward Hughes, Marc Lanctot, Thore Graepel
- A Recipe for Training Neural Networks, Andrej Karpathy
- Emotionally Adaptive Intelligent Tutoring Systems using POMDPs, Natasha Jaques
- Sapiens, Yuval Noah Harari
1 Listener
About TalkRL Podcast: All Reinforcement Learning, All the Time
TalkRL: The Reinforcement Learning Podcast
08/01/19 • 1 min
August 2, 2019
Transcript
The idea with TalkRL Podcast is to hear from brilliant folks from across the world of Reinforcement Learning, both research and applications. As much as possible, I want to hear from them in their own language. I try to get to know as much as I can about their work before hand.
And Im not here to convert anyone, I want to reach people who are already into RL. So we wont stop to explain what a value function is, for example. Though we also wont assume everyone has read the very latest papers.
Why am I doing this? Because it’s a great way to learn from the most inspiring people in the field! There’s so much happening in the universe of RL, and there’s tons of interesting angles and so many fascinating minds to learn from.
Now I know there is no shortage of books, papers, and lectures, but so much goes unsaid.
I mean I guess if you work at MILA or AMII or Vector Institute, you might be having these conversations over coffee all the time, but I live in a little village in the woods in BC, so for me, these remote interviews are like a great way to have these conversations, and I hope sharing with the community makes it more worthwhile for everyone.
In terms of format, the first 2 episodes were interviews in longer form, around an hour long. Going forward, some may be a lot shorter, it depends on the guest.
If you want want to be a guest or suggest a guest, goto talkrl.com/about, you will find a link to a suggestion form.
Thanks for listening!
1 Listener
Pierluca D'Oro and Martin Klissarov
TalkRL: The Reinforcement Learning Podcast
11/13/23 • 57 min
Pierluca D'Oro and Martin Klissarov on Motif and RLAIF, Noisy Neighborhoods and Return Landscapes, and more!
Pierluca D'Oro is PhD student at Mila and visiting researcher at Meta.
Martin Klissarov is a PhD student at Mila and McGill and research scientist intern at Meta.
Featured References
Motif: Intrinsic Motivation from Artificial Intelligence Feedback
Martin Klissarov*, Pierluca D'Oro*, Shagun Sodhani, Roberta Raileanu, Pierre-Luc Bacon, Pascal Vincent, Amy Zhang, Mikael Henaff
Policy Optimization in a Noisy Neighborhood: On Return Landscapes in Continuous Control
Nate Rahn*, Pierluca D'Oro*, Harley Wiltzer, Pierre-Luc Bacon, Marc G. Bellemare
To keep doing RL research, stop calling yourself an RL researcherPierluca D'Oro
Neurips 2024 RL meetup Hot takes: What sucks about RL?
TalkRL: The Reinforcement Learning Podcast
12/23/24 • 17 min
What do RL researchers complain about after hours at the bar? In this "Hot takes" episode, we find out!
Recorded at The Pearl in downtown Vancouver, during the RL meetup after a day of Neurips 2024.
Special thanks to "David Beckham" for the inspiration :)
Sven Mika
TalkRL: The Reinforcement Learning Podcast
08/19/22 • 34 min
Sven Mika is the Reinforcement Learning Team Lead at Anyscale, and lead committer of RLlib. He holds a PhD in biomathematics, bioinformatics, and computational biology from Witten/Herdecke University.
Featured References
RLlib Documentation: RLlib: Industry-Grade Reinforcement Learning
Ray: Documentation
RLlib: Abstractions for Distributed Reinforcement Learning
Eric Liang, Richard Liaw, Philipp Moritz, Robert Nishihara, Roy Fox, Ken Goldberg, Joseph E. Gonzalez, Michael I. Jordan, Ion Stoica
Episode sponsor: Anyscale
Ray Summit 2022 is coming to San Francisco on August 23-24.
Hear how teams at Dow, Verizon, Riot Games, and more are solving their RL challenges with Ray's RLlib.
Register at raysummit.org and use code RAYSUMMIT22RL for a further 25% off the already reduced prices.
Jessica Hamrick
TalkRL: The Reinforcement Learning Podcast
11/12/19 • 63 min
Dr. Jessica Hamrick is a Research Scientist at DeepMind. She holds a PhD in Psychology from UC Berkeley.
Featured References
Structured agents for physical construction
Victor Bapst, Alvaro Sanchez-Gonzalez, Carl Doersch, Kimberly L. Stachenfeld, Pushmeet Kohli, Peter W. Battaglia, Jessica B. Hamrick
Analogues of mental simulation and imagination in deep learning
Jessica Hamrick
Additional References
- Metacontrol for Adaptive Imagination-Based Optimization
Jessica B. Hamrick, Andrew J. Ballard, Razvan Pascanu, Oriol Vinyals, Nicolas Heess, Peter W. Battaglia - Surprising Negative Results for Generative Adversarial Tree Search
Kamyar Azizzadenesheli, Brandon Yang, Weitang Liu, Zachary C Lipton, Animashree Anandkumar - Metareasoning and Mental Simulation
Jessica B. Hamrick - Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm
David Silver, Thomas Hubert, Julian Schrittwieser, Ioannis Antonoglou, Matthew Lai, Arthur Guez, Marc Lanctot, Laurent Sifre, Dharshan Kumaran, Thore Graepel, Timothy Lillicrap, Karen Simonyan, Demis Hassabis - Object-oriented state editing for HRL
Victor Bapst, Alvaro Sanchez-Gonzalez, Omar Shams, Kimberly Stachenfeld, Peter W. Battaglia, Satinder Singh, Jessica B. Hamrick - FeUdal Networks for Hierarchical Reinforcement Learning
Alexander Sasha Vezhnevets, Simon Osindero, Tom Schaul, Nicolas Heess, Max Jaderberg, David Silver, Koray Kavukcuoglu - PILCO: A Model-Based and Data-Efficient Approach to Policy Search
Marc Peter Deisenroth, Carl Edward Rasmussen - Blueberry Earth
Anders Sandberg
David Silver @ RCL 2024
TalkRL: The Reinforcement Learning Podcast
08/26/24 • 11 min
David Silver is a principal research scientist at DeepMind and a professor at University College London.
This interview was recorded at UMass Amherst during RLC 2024.
References
- Discovering Reinforcement Learning Algorithms, Oh et al -- His keynote at RLC 2024 referred to more recent update to this work, yet to be published
- Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm, Silver et al 2017 -- the AlphaZero algo was used in his recent work on AlphaProof
- AlphaProof on the DeepMind blog
- AlphaFold on the DeepMind blog
- Reinforcement Learning Conference 2024
- David Silver on Google Scholar
Sharath Chandra Raparthy
TalkRL: The Reinforcement Learning Podcast
02/12/24 • 40 min
Sharath Chandra Raparthy on In-Context Learning for Sequential Decision Tasks, GFlowNets, and more!
Sharath Chandra Raparthy is an AI Resident at FAIR at Meta, and did his Master's at Mila.
Featured Reference
Generalization to New Sequential Decision Making Tasks with In-Context Learning
Sharath Chandra Raparthy , Eric Hambro, Robert Kirk , Mikael Henaff, , Roberta Raileanu
Additional References
- Sharath Chandra Raparthy Homepage
- Human-Timescale Adaptation in an Open-Ended Task Space, Adaptive Agent Team 2023
- Data Distributional Properties Drive Emergent In-Context Learning in Transformers, Chan et al 2022
- Decision Transformer: Reinforcement Learning via Sequence Modeling, Chen et al 2021
Natasha Jaques 2
TalkRL: The Reinforcement Learning Podcast
03/14/23 • 46 min
Hear about why OpenAI cites her work in RLHF and dialog models, approaches to rewards in RLHF, ChatGPT, Industry vs Academia, PsiPhi-Learning, AGI and more!
Dr Natasha Jaques is a Senior Research Scientist at Google Brain.
Featured References
Way Off-Policy Batch Deep Reinforcement Learning of Implicit Human Preferences in Dialog
Natasha Jaques, Asma Ghandeharioun, Judy Hanwen Shen, Craig Ferguson, Agata Lapedriza, Noah Jones, Shixiang Gu, Rosalind Picard
Sequence Tutor: Conservative Fine-Tuning of Sequence Generation Models with KL-control
Natasha Jaques, Shixiang Gu, Dzmitry Bahdanau, José Miguel Hernández-Lobato, Richard E. Turner, Douglas Eck
PsiPhi-Learning: Reinforcement Learning with Demonstrations using Successor Features and Inverse Temporal Difference Learning
Angelos Filos, Clare Lyle, Yarin Gal, Sergey Levine, Natasha Jaques, Gregory Farquhar
Basis for Intentions: Efficient Inverse Reinforcement Learning using Past Experience
Marwa Abdulhai, Natasha Jaques, Sergey Levine
Additional References
- Fine-Tuning Language Models from Human Preferences, Daniel M. Ziegler et al 2019
- Learning to summarize from human feedback, Nisan Stiennon et al 2020
- Training language models to follow instructions with human feedback, Long Ouyang et al 2022
Julian Togelius
TalkRL: The Reinforcement Learning Podcast
07/25/23 • 40 min
Julian Togelius is an Associate Professor of Computer Science and Engineering at NYU, and Cofounder and research director at modl.ai
Featured References
Choose Your Weapon: Survival Strategies for Depressed AI Academics
Julian Togelius, Georgios N. Yannakakis
Learning Controllable 3D Level Generators
Zehua Jiang, Sam Earle, Michael Cerny Green, Julian Togelius
PCGRL: Procedural Content Generation via Reinforcement Learning
Ahmed Khalifa, Philip Bontrager, Sam Earle, Julian Togelius
Illuminating Generalization in Deep Reinforcement Learning through Procedural Level Generation
Niels Justesen, Ruben Rodriguez Torrado, Philip Bontrager, Ahmed Khalifa, Julian Togelius, Sebastian Risi
Show more best episodes
Show more best episodes
FAQ
How many episodes does TalkRL: The Reinforcement Learning Podcast have?
TalkRL: The Reinforcement Learning Podcast currently has 62 episodes available.
What topics does TalkRL: The Reinforcement Learning Podcast cover?
The podcast is about Podcasts, Technology, Artificial Intelligence and Machine Learning.
What is the most popular episode on TalkRL: The Reinforcement Learning Podcast?
The episode title 'Natasha Jaques' is the most popular.
What is the average episode length on TalkRL: The Reinforcement Learning Podcast?
The average episode length on TalkRL: The Reinforcement Learning Podcast is 53 minutes.
How often are episodes of TalkRL: The Reinforcement Learning Podcast released?
Episodes of TalkRL: The Reinforcement Learning Podcast are typically released every 19 days, 2 hours.
When was the first episode of TalkRL: The Reinforcement Learning Podcast?
The first episode of TalkRL: The Reinforcement Learning Podcast was released on Aug 1, 2019.
Show more FAQ
Show more FAQ