Show HN: I analyzed why my post got 0 votes and built this
Continuous scanning: Analyzes all new HN posts from accounts with <100 karma, scoring them for technical merit, originality, and problem-solving value AI curation: Every 12 hours, an LLM deep-dives into the top 10 candidates, checking GitHub repos, documentation quality, and actual utility
The result is what you see at the link - a curated list of overlooked quality posts that deserve more attention. The interesting part: I barely wrote any criteria. I just told Claude "open source good, pure commercial bad, working demos good" and let it figure out the scoring. The AI assessment varies slightly each run, which actually makes it more interesting. GitHub: https://github.com/DG1001/hn-gems
Is this useful? Do you have ideas how to improve this tool if necessary?
(And yes, my XaresAICoder that got 0 votes? The AI thinks it's actually pretty good. I'll take that as a win.)
Take a look at /new, choked with AI products and “Ask HN: How come Vibe Coding Doesn’t Work For Me Although It Works For Everyone Else?” Or “Show HN: I Vibe Coded Something and It Almost Works”, “Is MCP really that evil program from Tron?”, etc.
The cheat code for August 2025 is it’s not about AI and if it an AI-oriented product it doesn’t say so right out front.