Inside the Belly of the Beast: How LLMs Work

Jeff Prosise

Wintellect Cofounder and InRule CAIO

About the session

In 2017, a landmark paper entitled “Attention is All You Need” turned the deep-learning world on its head and laid the foundation for today’s Large Language Models. In the paper, we learned about the transformer architecture and its unparalleled ability to understand, and even to generate, human language. Yet the inner workers of transformers are a mystery to most.

This session takes a deep dive into natural-language processing, starting with simple word embeddings and text classifiers, moving to Recurrent Neural Networks (RNNs), and ultimately landing on transformers.

You’ll come away understanding what a transformer is, how it works, and what happens as words flow through a transformer and are converted from static embeddings with low information density into rich, context-aware embeddings that capture the meaning of the words around them. You’ll also learn what context windows really are and why GPUs are so crucial to a transformer's operation.

It will give you a new appreciation for the brilliance of the transformer architecture. And it might make you a better engineer.

Get inspiration & news from us

I agree that Cornerstone will send me news via e-mail