- We build the Thread Inference Model (TIM) based on the transformer architecture, and its dedicated runtime TIMRUN.
- TIM + TIMRUN = Intelligent workflow generation, context engineering, and multi-hop tool use happens at the runtime level
- TIM + TIMRUN supports virtually unlimited reasoning enabled by context pruning, significantly improves the efficiency for long-horizon reasoning tasks
I’m more than ready to stop my workflow being derailed by opening new chats—looking forward to LLM assistants that actually remember our conversation.
kevin8704 · 5h ago
Really cool work! I saw the “Selective Working Memory” section, are there hooks to swap in a custom retrieval store or memory layer, or is that all internal to TIM?
- TIM + TIMRUN = Intelligent workflow generation, context engineering, and multi-hop tool use happens at the runtime level
- TIM + TIMRUN supports virtually unlimited reasoning enabled by context pruning, significantly improves the efficiency for long-horizon reasoning tasks
- Inference API is live at https://subconscious.dev/
- More details: https://github.com/subconscious-systems/TIMRUN
Thanks for sharing!