Coffee Time Papers
Reading AI papers while enjoying a cup of coffee
May 27, 2024
--
- Attention is All You Need
- Leave No Context Behind
- Scaling Laws for Neural Language Models
- Leave No Context Behind
- Kolmogorov–Arnold Networks
- The Platonic Representation Hypothesis
- Your Transformer is Secretly Linear
- Demystifying Embedding Spaces Using LLMs
- Take it to the Limit: Peak Prediction-driven Resource Overcommitment in Datacenters
- Long-term Forecasting with Time-series Dense Encoder
- Mixture of Agents
- Can LLMs Learn By Teaching
- Google Vizier: A Service for Black-Box Optimization
- Mixture of Depths
- Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets
- MiniMalloc: A Lightweight Memory Allocator for Hardware-Accelerated Machine Learning