How LLMs Work

David Banks Speaker
Duke University
 
Tuesday, Aug 6: 8:30 AM - 10:20 AM
Introductory Overview Lectures 
Oregon Convention Center 
LLMs are deep neural networks. This talk describes the relationships between deep recurrent neural networks, transformers, attention, and token embedding that underlie large language models. Combined with reinforcement learning, large language models produce chatbots that show powerful performance on many axes, including text and image generation. This talk also discusses practical issues in training them.