CS224U
Chris Potts
1 Listener
All episodes
Best episodes
Top 10 CS224U Episodes
Goodpods has curated a list of the 10 best CS224U episodes, ranked by the number of listens and likes each episode have garnered from our listeners. If you are listening to CS224U for the first time, there's no better place to start than with one of these standout episodes. If you are a fan of the show, vote for your favorite CS224U episode by adding your comments to the episode page.
02/23/23 • 86 min
Lessons learned about benchmarking, adversarial testing, the dangers of over- and under-claiming, and AI alignment.
Transcript: https://web.stanford.edu/class/cs224u/podcast/bowman/
- Sam's website
- Sam on Twitter
- NYU Linguistics
- NYU Data Science
- NYU Computer Science
- Anthropic
- SNLI paper: A large annotated corpus for learning natural language inference
- SNLI leaderboard
- FraCaS
- SICK
- A SICK cure for the evaluation of compositional distributional semantic models
- SemEval-2014 Task 1: Evaluation of Compositional Distributional Semantic Models on Full Sentences through Semantic Relatedness and Textual Entailment
- RTE Knowledge Resources
- Richard Socher
- Chris Manning
- Andrew Ng
- Ray Kurtzweil
- SQuAD
- Gabor Angeli
- Adina Williams
- Adina Williams podcast episode
- MultiNLI paper: A broad-coverage challenge corpus for sentence understanding through inference
- MultiNLI leaderboards
- Twitter discussion of LLMs and negation
- GLUE
- SuperGLUE
- DecaNLP
- GPT-3 paper: Language Models are Few-Shot Learners
- FLAN
- Winograd schema challenges
- BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
- JSALT: General-Purpose Sentence Representation Learning
- Ellie Pavlick
- Ellie Pavlick podcast episode
- Tal Linzen
- Ian Tenney
- Dipanjan Das
- Yoav Goldberg
- Fine-grained Analysis of Sentence Embeddings Using Auxiliary Prediction Tasks
- Big Bench
- Upwork
- Surge AI
- Dynabench
- Douwe Kiela
- Douwe Kiela podcast episode
- Ethan Perez
- NYU Alignment Research Group
- Eliezer Shlomo Yudkowsky
- Alignment Research Center
- Redwood Research
- Percy Liang podcast episode
- Richard Socher podcast episode
1 Listener
11/07/22 • 68 min
Leaving Ohio, being back in Belgium, organizing NAACL 2022, reviewing at NLP-scale, universal dependencies, and doing NLU before it was cool.
Transcript: https://web.stanford.edu/class/cs224u/podcast/demarneffe/
- Marie's website
- Generating Typed Dependency Parses from Phrase Structure Parses
- Universal Dependencies project
- OSU Linguistics
- NAACL 2022
- Dan Jurafsky
- Dan Roth
- Chris Manning
- ARR
- Priscilla Rasmussen
- Transactions of the ACL
- Finding Contradictions in Text
- Not a simple yes or no: Uncertainty in indirect answers
- Recognizing Textual Entailment
- Anna Rafferty
- Scott Grimm
- "Was It Good? It Was Provocative." Learning the Meaning of Scalar Adjectives
- Did It Happen? The Pragmatic Complexity of Veridicality Assessment
- Yejin Choi
- Yejin Choi's ACl 2022 talk
- Barbara Plank
- Linguistically debatable or just plain wrong?
- Jesse Dodge
- Reproducibility badges at NAACL 2022
- Stanford Sentiment Treebank
- Judith Tonhauser
- Nan-Jiang Jiang
- Lauri Karttunen
- Climbing towards NLU: On Meaning, Form, and Understanding in the Age of Data
- Microsoft DeBERTa surpasses human performance on the SuperGLUE benchmark
- Daniel Zeman
- Marta Recasens
06/27/22 • 86 min
Birth narratives, stable static representations, NLP for everyone, AI2 and Semantic Scholar, the mission of Ukrainian Catholic University, and books books books.
Transcript: https://web.stanford.edu/class/cs224u/podcast/antoniak/
- Maria's website
- Maria on Twitter
- Semantic Scholar
- Elliott Ash
- ETH Zurich Center for Law and Economics
- Text As Data (TADA) 2022
- David Mimno
- A computational reading of a birth stories community
- r/BabyBumps
- Roger Shank
- Nate Chambers
- ICWSM 2022 workshop: BERT for Social Sciences and Humanities
- Measuring Word Similarity with BERT (Sephora Makeup Reviews)
- Melanie Walsh
- word2vec
- BERT
- Nick Vincent's Twitter thread on Meta's OPT-175B filtering strategies
- Stemming
- Alexandra Schofield
- LDA
- LSA
- GloVe
- Evaluating the stability of embedding-based word similarities
- Narrative datasets through the lenses of NLP and HCI
- Belmont report
- Casey Fiesler
- Naive Bayes
- Allen Institute
- CORD-19 dataset, which appeared March 16, 2020!
- Books books books
- Pushkin Press
- New York Review Books
- Posthumous Memoirs of Brás Cubas
- And Then There Were None
- Stanisław Lem
- Jeff VanderMeer
- Italo Calvino
- Jorge Luis Borges
- xkcd
- War and Peace
- Middlemarch
- Beloved
- Novelist Cormac McCarthy's tips on how to write a great science paper
- Blood Meridian
- No Country for Old Men (book)
- No Country for Old Men (movie)
- The Road
- Talking a visual walk through Burnt Norton
- Ukrainian Catholic University
- Support Ukraine Now: Real Ways You can H...
04/18/22 • 81 min
Hugging Face, multimodality, data and model auditing, ethics review, adversarial testing, attention as more and less than you ever needed, neural information retrieval, philosophy of mind and consciousness, augmenting human creativity, openness in science, and a defininitive guide to pronouncing Douwe.
Transcript: https://web.stanford.edu/class/cs224u/podcast/kiela/
- Douwe's website
- Hugging Face
- Grounding semantics in olfactory perception
- Model Cards for Model reporting
- Datasheets for datasets
- Dynabench
- Hugging Face Spaces
- http://www.isattentionallyouneed.com
- The Annotated S4
- Retrieval-Augmented Generation for knowledge-intensive NLP tasks
- Language models as slightly consciousness
- Fields of wheat as slightly pasta
- True few-shot learning with language models
- https://believermag.com/ghosts/
- I Am A Strange Loop
- AI Dungeon
- LIGHT
- Good first issue
04/25/22 • 85 min
Pronouncing "ColBERT", the origins of ColBERT, doing NLP from an IR perspective, how getting "scooped" can be productive, OpenQA and related tasks, PhD journeys, why even retrieval plus attention is not all you need, multilingual knowledge-intensive NLP, and aiming high in research projects.
Transcript: https://web.stanford.edu/class/cs224u/podcast/khattab/
- Omar's website
- Matei Zaharia
- Keshav Santhanam
- Steven Colbert thowing paper with Obama
- The ColBERT paper and the ColBERTv2 paper
- DeepImpact: Learning passage impacts for inverted indexes
- DPR: Dense passage retrieval for open-domain question answering
- Incorporating query term independence assumption for efficient retrieval and ranking using deep neural networks
- DeepCT: Context-aware sentence/passage term importance estimation for first stage retrieval
- Reading Wikipedia to answer open-domain questions
- ORQA: Latent retrieval for weakly supervised open domain question answering
- QRECC
- ColBERT-QA: Relevance-guided Supervision for OpenQA with ColBERT
- Baleen: Robust Multi-Hop Reasoning at Scale via Condensed Retrieval
- Passage reranking with BERT
- UniK-QA: Unified Representations of Structured and Unstructured Knowledge for Open-Domain Question Answering
- Self-driving search engines: The neural hype and comparisons against weak baselines
- Mohammad Hammoud
- RAG: Retrieval-augmented generation for knowledge-intensive NLP tasks
- Hindsight: Posterior-guided training of retrievers for improved open-ended generation
- Learning Cross-Lingual IR from an English Retriever
- Blog post: A moderate proposal for radically better AI-powered Web search
- Blog post: Building scalable, explainable, and adaptive NLP models with retrieval
- XOR-TyDi
05/09/22 • 83 min
Grounding through pure language modeling objectives, the origins or probing, the nature of understanding, the future of system assessment, signs of meaningful progress in the field, and having faith in yourself.
Transcript: https://web.stanford.edu/class/cs224u/podcast/pavlick/
- Ellie's website
- The LUNAR Lab
- MIT Scientist Captures 90,000 Hours of Video of His Son’s First Words, Graphs It
- Michael Frank
- Spot robots
- Dylan Ebert
- Ian Tenney
- What do you learn from context? Probing for sentence structure in contextualized word representations
- BERT Rediscovers the Classical NLP Pipeline
- JSALT: General-Purpose Sentence Representation Learning
- Sam Bowman
- Skip thought vectors
- What you can cram into a single $&!#* vector: Probing sentence embeddings for linguistic properties
- Hex
- Charlie Lovering
- Designing and interpreting probes with control tasks
- Jerry Fodor
- Been Kim
- Mycal Tucker
- What if this modified that? Syntactic interventions via counterfactual embeddings
- Yonatan Belinkov
- HANS: Right for the Wrong Reasons: Diagnosing Syntactic Heuristics in Natural Language Inference
- Conceptual pacts and lexical choice in conversation
- Locating and editing factual knowledge in GPT
- Could a purely self-supervised language model achieve grounded language understanding?
- Dartmouth Summer Research Project on Artificial Intelligence (1956)
- Equal numbers of neuronal and nonneuronal cells make the human brain an isometrically scaled-up primate brain
05/16/22 • 83 min
Coast-to-coast professional journeys, multilingual NLP, teaching in a fast-changing field, the history of hate speech detection in NLP, ethics review of NLP research, research on sensitive topics, mentoring researchers, and optimizing for your own passions.
Transcript: https://web.stanford.edu/class/cs224u/podcast/tsvetkov/
- Yulia's website
- TsvetShop
- Shuly Wintner
- Just when I thought I was out ...
- Algorithms for NLP
- HMMs
- Kneser–Ney smoothing
- Noah Smith
- Demoting racial bias in hate speech detection
- The risk of racial bias in hate speech detection
- Demoting racial bias in hate speech detection
- Fortifying toxic speech detectors against veiled toxicity
- This is the daily stormer's playbook
- Microaggressions.com
- Finding microaggressions in the wild: A case for locating elusive phenomena in social media posts
- https://delphi.allenai.org
- Delphi: Towards Machine Ethics and Norms
- Yejin Choi
06/13/22 • 87 min
Realizing that Foundation Models are a big deal, scaling, why Percy founded CRFM, Stanford's position in the field, benchmarking, privacy, and CRFM's first and next 30 years.
Transcript: https://web.stanford.edu/class/cs224u/podcast/liang/
- Percy's website
- Percy on Twitter
- CRFM
- On the opportunities and risks of foundation models
- ELMo: Deep contextualized word representations
- BERT: Pre-training of deep bidirectional Transformers for language understanding
- Sam Bowman
- GPT-2
- Adversarial examples for evaluating reading comprehension systems
- System 1 and System 2
- The Unreasonable Effectiveness of Data
- Chinchilla: Training Compute-Optimal Large Language Models
- GitHub Copilot
- LaMDA: Language models for dialog applications
- AI Test Kitchen
- DALL-E 2
- Richer Socher on the CS224U podcast
- you.com
- Chris Ré
- Fei-Fei Li
- Chris Manning
- HAI
- Rob Reich
- Erik Brynjolfsson
- Dan Ho
- Russ Altman
- Jeff Hancock
- The time is now to develop community norms for the release of foundation models
- Twitter Spaces event
- Best practices for deploying language models
- Model Cards for model reporting
- Datasheets for datasets
- Strathern's law
10/04/22 • 82 min
Coding puzzles, practices, and education, structured prediction, the culture of Hugging Face, large models, and the energy of New York.
Transcript: https://web.stanford.edu/class/cs224u/podcast/rush/
- Sasha's website
- Sasha on Twitter
- Sasha on the Humans of AI podcast
- Sasha on The Thesis Review Podcast with Sean Welleck
- Sasha on the Talking Machines Podcast
- Sasha interviewed by Sayak Paul
- Hugging Face
- PyTorch
- The Annotated Transformer
- The Annotated Alice
- The Annotated S4
- Sasha and Dan Oneață's declarative graphics library Chalk
- Drawing Big Ben in Chalk
- OpenNMT
- Ken Shan
- Blog post by Ken and Dylan Thurston
- Edward Z. Yang
- Stuart Shieber
- Literate programming
- Soumith Chintala
- Lua Torch
- TensorFlow
- Graham Neubig
- Chris Dyer
- DyNet
- JAX
- jax.vmap
- Matt Johnson
- Finale Doshi-Velez, whose undergrad ML course inspired and informed Sasha's
- Tensor Puzzles
- GPU Puzzles
- A tweet that Chris added to his CV
- Adam Paszke
- Dougal MacLaurin
- Dex
- Named Tensor notation
- Named Tensors in PyTorch
- TorchDim
- Mini Torch
- Torch-Struct
- Sarah Hooker's paper 'The hardware lottery'
- Jacob Andreas
- Kevin Ellis
- Hugging Face transformers library
- Hugging Face datasets library
- Hugging Face diffusers library
- Hugging Face evaluate library
- scikit-learn
- Big Science blog
- BLOOM
- The Technology Behind BLOOM Training
- CRFM
- Eleuther
- T0 and PromptSource
- Washington Post: Big Tech builds AI with bad data. So scientists sought better data
01/27/23 • 88 min
AI and social science, the causal revolution in economics, predictions about the impact of AI, teaching MBAs, productizing AI, and a journey from Tel Aviv to Princeton to Stanford.
Transcript: https://web.stanford.edu/class/cs224u/podcast/goldberg/
- Amir's website
- Amir on Twitter
- Computational Culture Lab
- ChatGPT
- Laura Nelson
- Bart Bonikowski
- Chris Winship
- Bernie Koch
- Treebanks
- BIG-bench
- Guido Imbens
- Endogeneity
- Susan Athey
- Cambridge Analytica
- Prediction Machines
- Speech and Language Processing
- DALL-E 2
- Midjourney
- Stable Diffusion
- Postmodernism, or, the Cultural Logic of Late Capitalism
- Turing test
- Matt Salganik
- Paul DiMaggio
Show more best episodes
Show more best episodes
FAQ
How many episodes does CS224U have?
CS224U currently has 16 episodes available.
What topics does CS224U cover?
The podcast is about Podcasts and Technology.
What is the most popular episode on CS224U?
The episode title 'Sam Bowman on benchmarking and AI alignment' is the most popular.
What is the average episode length on CS224U?
The average episode length on CS224U is 85 minutes.
How often are episodes of CS224U released?
Episodes of CS224U are typically released every 7 days, 4 hours.
When was the first episode of CS224U?
The first episode of CS224U was released on Apr 11, 2022.
Show more FAQ
Show more FAQ