#transformer

1 posts

LLM Fundamentals: Parameters, Embeddings, and Attention

How LLM parameters encode meaning, what embedding dimensions actually represent, and why Transformer attention is computed in parallel.