How Common Is Multiple Invention? (construction-physics.com)
22 points by rbanffy 5h ago 12 comments
LLMs and Elixir: Windfall or Deathblow? (zachdaniel.dev)
230 points by uxcolumbo 1d ago 119 comments
Tokasaurus: An LLM Inference Engine for High-Throughput Workloads
113 rsehrlich 9 6/5/2025, 9:27:07 PM scalingintelligence.stanford.edu ↗
https://github.com/ScalingIntelligence/tokasaurus/blob/65efb...
I’m honestly impressed that a pure python implementation can beat out vLLM and SGLang. Granted they lean on FlashInfer, and of course torch.compile has gotten incredibly powerful in the last few years. Though dynamic shapes have still been a huge thorn in my side, I’ll need to look closer at how they pulled it off…
Because Tokasaurus was mentioned as better than Ollama for conducting darwinian godel machine operations (self-improvement), I looked for the linked repo on GitHub and it was 404. So glad it is back https://github.com/ScalingIntelligence/tokasaurus.
Looks like they don't compare to TensorRT-LLM throughput numbers which, last I checked, are SOTA in open source.
Also, this seems very useful for generating synthetic data or labelling a bunch of data. 6k batch size is small for data labelling.