Session
The math behind LLMs: GPTs from scratch
Large Language Models can seem magics, but in fact they are built on layers of math, research, and innovations over the years. In this session, we'll cover the math behind them and see how you could build one if you had unlimited time, money, and training data.
We'll start with a simple bigram prediction model and work our way up. We'll cover neural network such as backpropagation and gradient descent. By the end of this talk, you'll see how you could build a very tiny GPT.
Eugene Meidinger
Data training that respects your time
Pittsburgh, Pennsylvania, United States
Links
Please note that Sessionize is not responsible for the accuracy or validity of the data provided by speakers. If you suspect this profile to be fake or spam, please let us know.
Jump to top