Cover image

Greedy Enough to Win: When Loss Starts Driving the Learning Rate

Opening — Why this matters now Modern deep learning training is an odd contradiction. We obsess over architectures, data curation, and trillion-token scaling laws—then quietly accept Cosine Annealing as if it were gravity. Learning rate schedules are often inherited, not argued for. This paper challenges that complacency with a scheduler that does something almost offensive in its simplicity: it just watches the loss and reacts. ...

December 17, 2025 · 3 min · Zelina