Ingest Log

Entries are appended chronologically as sources are ingested.

[2026-04-11] ingest | An Image is Worth 16x16 Words (ViT)

[2026-04-10] ingest | Training language models to follow instructions with human feedback (InstructGPT)

[2026-04-09] ingest | CLIP: Learning Transferable Visual Models From Natural Language Supervision

[2026-04-08] ingest | Emergent Abilities of Large Language Models

[2026-04-06] ingest | RoPE: Enhanced Transformer with Rotary Position Embedding

[2026-04-04] ingest | Direct Preference Optimization: Your Language Model is Secretly a Reward Model

[2026-04-05] ingest | Attention Is All You Need

[2026-04-05] ingest | LoRA: Low-Rank Adaptation of Large Language Models

[2026-04-05] ingest | FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness

[2026-04-05] ingest | Efficient Memory Management for Large Language Model Serving with PagedAttention

[2026-04-05] ingest | Chain-of-Thought Prompting Elicits Reasoning in Large Language Models

[2026-04-05] ingest | Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets

[2026-04-05] ingest | A Systematic Empirical Study of Grokking: Depth, Architecture, Activation, and Regularization

[2026-04-05] ingest | Mamba: Linear-Time Sequence Modeling with Selective State Spaces

[2026-04-05] ingest | Fast Inference from Transformers via Speculative Decoding

[2026-04-05] ingest | Falcon Perception: Early-Fusion Transformer for Open-Vocabulary Grounding and Segmentation