LLMs as Retrieval and Recommendation Engines
3 pongogogo 2 9/12/2025, 5:28:01 PM medium.com ↗
Comments (2)
taintech · 4h ago
This is a cool idea, but the cost is a killer. Normally, you would run recommendations and pre-cache them for the users or items they are associated with. Running a giant LLM for every user's recommendation is thousands of times more expensive and slower than current methods. It just doesn't seem practical for a large number of users.
pongogogo · 4h ago
The post mentions an approach of using a large model to generate labels and then distilling this into a smaller model to lower cost (though it doesn't provide an example)