Pretraining a LLM with less than $50 budget which outperforms Google BERT

6 mrxhacker99 1 9/1/2025, 8:02:32 PM medium.com ↗

Comments (1)

spindump8930 · 3h ago
The title makes it sound nice but the reported results are worse than random baselines on several benchmarks, including ones to claim superiority over BERT. At a glance, Hellaswag, boolq, winogrande are all at or below random guessing. At best this is a fun model with broken evaluation. At worst this is medium spam for clout farming - which won't work on anyone who can read the tables.