Ask HN: Why is LLM training still GPU-hungry despite DeepSeek?

1 takinola 0 7/15/2025, 2:25:19 PM
When DeepSeek released R-1 everyone thought that signaled the end of the GPU-intensive LLM training approach. It does not appear to have worked out that way as GPU demand continues to grow unabated. What happened? Is the DeepSeek training method unreproducible or impractical in some way?

Comments (0)

No comments yet