LLM Notes
LLM Notes
Links:
Tokenization (
BPE, WordPiece, SentencePiece):Attention:
Positional Encoding:
Normalization:
Activation Functions & attention inference trick:
Mixture of Experts (MoE):
LLM tricks & advanced techniques:
Finetuning:
Links:
Tokenization (BPE, WordPiece, SentencePiece):
Attention:
Positional Encoding:
Normalization:
Activation Functions & attention inference trick:
Mixture of Experts (MoE):
LLM tricks & advanced techniques:
Finetuning: