LLM Fundamentals: Parameters, Embeddings, and Attention
How LLM parameters encode meaning, what embedding dimensions actually represent, and why Transformer attention is computed in parallel.
1 posts
How LLM parameters encode meaning, what embedding dimensions actually represent, and why Transformer attention is computed in parallel.