key topics to learn how llms work, all it takes is < 2 years if you have cs foundation > tokenization and embeddings
> positional embeddings (absolute, rope, alibi)
> self attention and multihead attention
> transformers
> qkv
> sampling params: temperature, top-k top-p
> kv
Essential LLM Concepts: Tokenization, Attention, and Sampling
By
–
Leave a Reply