End-to-end long-horizon reasoning with one Transformer model

5 hyluo 3 7/23/2025, 1:15:39 PM subconscious.dev ↗

Comments (3)

hyluo · 6h ago
- We build the Thread Inference Model (TIM) based on the transformer architecture, and its dedicated runtime TIMRUN.

- TIM + TIMRUN = Intelligent workflow generation, context engineering, and multi-hop tool use happens at the runtime level

- TIM + TIMRUN supports virtually unlimited reasoning enabled by context pruning, significantly improves the efficiency for long-horizon reasoning tasks

- Inference API is live at https://subconscious.dev/

- More details: https://github.com/subconscious-systems/TIMRUN

datapalo · 4h ago
I’m more than ready to stop my workflow being derailed by opening new chats—looking forward to LLM assistants that actually remember our conversation.
kevin8704 · 5h ago
Really cool work! I saw the “Selective Working Memory” section, are there hooks to swap in a custom retrieval store or memory layer, or is that all internal to TIM?

Thanks for sharing!