This video kinda captures all the foundational AI Papers
1. Attention Is All You Need (2017) - Transformer architecture with self-attention for parallel training; foundation of all modern LLMs
2. GPT-3 (2020) - Demonstrated in-context learning at scale; models can learn tasks from prompts without fine-tuning
3. InstructGPT (2022) - RLHF alignment technique; smaller aligned models outperform larger unaligned ones
4. LoRA (2021) - Low-rank adapters for efficient fine-tuning; 10,000x fewer parameters, enables single-GPU training
5. RAG (2020) - Retrieval-Augmented Generation; models access external data to reduce hallucination and outdated knowledge
6. LLM-Based Agents Survey (2023) - Framework for AI agents with brain (planning), perception (context), and action (tools) components
7. Switch Transformers (2021) - Mixture of Experts architecture; trillion-parameter models with sparse activation for efficiency
8. DistilBERT (2019) - Knowledge distillation for compression; 40% smaller, 60% faster, 97% performance retention
9. LLM.int8() (2022) - Outlier-aware quantization; halves memory with 8-bit storage while preserving accuracy
10. Model Context Protocol (2024) - Anthropic's open standard for connecting models to tools, databases, and APIs seamlessly
Just Google them and get everything on Arxiv :)
Nov 8, 2025 · 8:44 AM UTC

