Language Understanding and LLMs with Christopher Manning
EPISODE 686
|
MAY
27,
2024
Watch
Follow
Share
About this Episode
Today, we're joined by Christopher Manning, the Thomas M. Siebel professor in Machine Learning at Stanford University and a recent recipient of the 2024 IEEE John von Neumann medal. In our conversation with Chris, we discuss his contributions to foundational research areas in NLP, including word embeddings and attention. We explore his perspectives on the intersection of linguistics and large language models, their ability to learn human language structures, and their potential to teach us about human language acquisition. We also dig into the concept of “intelligence” in language models, as well as the reasoning capabilities of LLMs. Finally, Chris shares his current research interests, alternative architectures he anticipates emerging beyond the LLM, and opportunities ahead in AI research.
About the Guest
Christopher Manning
Machine Learning in the Departments of Linguistics and Computer Science at Stanford University; Stanford Artificial Intelligence Laboratory (SAIL); Stanford NLP group; Stanford Institute for Human-Centered Artificial Intelligence (HAI)
Resources
- GloVe: Global Vectors for Word Representation
- 2024 IEEE John Von Neumann Medal: Christopher D. Manning
- Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank
- The Stanford Natural Language Inference (SNLI) Corpus
- A large annotated corpus for learning natural language inference
- Preference Tuning LLMs with Direct Preference Optimization Methods
- Emergent linguistic structure in artificial neural networks trained by self-supervision
- https://universaldependencies.org/
- The Case for Universal Dependencies
- Noam Chomsky website
- https://x.com/chrmanning/status/1768291977400103064
- Attention Is All You Need
- Direct Preference Optimization: Your Language Model is Secretly a Reward Model
- Proximal Policy Optimization Algorithms
- Pushdown Layers: Encoding Recursive Structure in Transformer Language Models
- Effective Approaches to Attention-based Neural Machine Translation
- Contrastive Learning of Medical Visual Representations from Paired Images and Text
- ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators

