Show HN: AuraCoder – Gen AI Learning Platform (auracoder.com)
2 points by zsmith99 5h ago 0 comments
Show HN: I'm 13 and I built an AI PDF Reader (github.com)
5 points by pro-grammer 21h ago 1 comments
Reinforcement Pre-Training
66 frozenseven 18 6/10/2025, 5:30:22 AM arxiv.org ↗
Let's assume this is a paradigm shift on the scale of Transformers / `Attention is all you need`. Companies build out new models and pump another $100 Billion through it. And then a year from now, another innovation comes out. Same circus. And again.
No one wants to be left behind but trying to keep up will sink smaller companies.
Yes, the more recent generation of GPUs optimize for attention math. But they are still fairly "general-purpose" accelerators as well. So when I see papers like this (interesting idea, btw!), my mental model for costs suggests that the CapEx to buy up the GPUs and build out the data centers would get re-used for this and 100s of other ideas and experiments.
And then the hope is that the best ideas will occupy more of the available capacity...
The downside is that this is going to be extremely expensive, so the data set to conduct RL will need to be curated.
[1]: https://news.ycombinator.com/item?id=41776324