Introduction
Natural Language Processing
1.
LLMs
1.1.
Architecture
1.1.1.
FeedForward
1.1.2.
Attention
1.1.3.
Transformer
1.1.4.
Mixture of Experts
1.1.5.
Encoders
1.1.6.
Decoders
1.1.7.
Encoder-Decoder
1.1.8.
Multi-Latent Attention
1.2.
Prompting
1.2.1.
Prompt Engineering
1.2.2.
In-Context Learning
1.2.3.
Few-Shot Learning
1.2.4.
Chain of Thought
1.2.5.
Tree of Thought
1.2.6.
Soft prompts
1.2.7.
Hard prompts
1.3.
Fine-tuning
1.3.1.
Supervised Fine-Tuning
1.3.2.
RLHF
1.3.3.
DPO
1.3.4.
GRPO
1.3.5.
PEFT
1.3.6.
LoRA
1.3.7.
QLoRA
1.3.8.
DoRA
1.3.9.
YaRN
1.4.
Agents
1.4.1.
Tool Use
1.4.2.
Reflection
1.4.3.
Multi Agent
1.4.4.
Planning
1.5.
RAG
1.5.1.
Chunks
1.5.2.
Sliding Window
1.5.3.
Graph RAG
1.6.
Model Compression
1.6.1.
Distillation
1.6.2.
Quantization
1.7.
Efficient Inference
1.7.1.
Fast Attention
1.7.2.
KV Cache
1.8.
Decoding
1.8.1.
Multi-Token Prediction
1.8.2.
Top-k
1.8.3.
Greedy
1.8.4.
Speculative
1.9.
Miscellaneous
1.9.1.
Rejection Sampling
1.9.2.
Emergent
1.9.3.
LLM As Judge
2.
Notable Models
2.1.
BERT
2.2.
Llama-3
2.3.
DeepSeek-R1
2.4.
DeepSeek-v3
2.5.
Qwen2.5
Evaluation
3.
Metrics
3.1.
Rouge
3.2.
Bleu
3.3.
pass@k