Generative AI - math from scratch!
Data Science/AI Researcher @ Shell | Founder @ Petroleum From Scratch | Scientific ML | Math for GenAI
📍Regression Math to LLM Math - Connecting the dots. Every model at the core of it is just y = f(X) and the f keeps getting prettier. 1. f being a linear vector of weights and X being row wise inputs, you’re in the Linear Regression regime. 2. f being a recursive matrix multiplication of weight matrices and sandwiched by non linear activation functions, X being structured matrix inputs, you’re in the deep neural network regime. 3. f being a transformer stack with key query and value interaction formulations in self and cross attention modes with other ingredients like position encoding and normalizations followed by a softmax, you’re in the GenAI LLM modelling regime with a multi class classification way of things i.e. the model classifying the output class from the vocab space. Pardon me for the oversimplification. But I hope it makes sense for some. Happy Learning! No hash-tags.