Show HN: JavaFactory – IntelliJ plugin to generate Java code

38 javafactory 12 5/20/2025, 11:29:45 AM github.com ↗
Hi HN,

I built a code generator plugin for IntelliJ that uses LLMs to create repetitive Java code like implementations, tests, and fixtures — based on custom natural-language patterns and annotation-based references.

Most tools like Copilot or Cursor aim to be general, but fail to produce code that actually fits a project structure or passes tests.

So I made something more explicit: define patterns + reference scope, and generate code consistently.

In this demo, 400 lines of Java were generated in 20 seconds — and all tests passed: https://www.youtube.com/watch?v=ReBCXKOpW3M

GitHub: https://github.com/JavaFactoryPluginDev/javafactory-plugin

Comments (12)

AugustoCAS · 5h ago
A side comment, I have found that configuring a few live templates in IntelliJ helps me to write a lot of the repetitive code just a handful of keystrokes regardless of the language.

Structural refactoring is another amazing feature that is worth knowing.

geodel · 9h ago
Feels very Java like. Factories, repositories, utils, patterns etc. Good stuff.
asdffdasy · 8h ago
yoDawgMemesFactory
simpaticoder · 8h ago
If the trend continues a program will look like "JavaFactory("<prompt>").compile().run();".
winrid · 7h ago
I've always wondered how long until we reach this. If every pc can run models locally, with a given seed and prompt it could be the ultimate compression. It's also hilarious.
imhoguy · 7h ago
Although very lossy compression, each invocation will be different, so that will inevitably circle back to "strong-static-LLM" prompts. What? wait..!
woodrowbarlow · 6h ago
LLMs at their core do produce reproducible results with a given seed. it's all the workflow stuff people do on top that tends to break reproducibility.
dibujaron · 1h ago
This is not the case for LLMs running on GPUs (which is most of them); GPUs are non-deterministic for this use-case due to the floating point math involved. there is no way to get perfectly deterministic output from OpenAI despite the presence of seed and temperature parameters.
likis · 8h ago
What LLM is it using? Is it something local? Or does it call out? It wasn't obvious from the docs, and I didn't want to dig through all of the code to figure it out. Should probably be clearly stated on the front page.

But the project looks interesting, I have been looking for something similar.

trollied · 8h ago
It uses openai.
p0w3n3d · 4h ago
As a programmer I feel bad if tests don't fail at the first run... It might show that they are not testing...
cess11 · 9h ago
The guide is a 404.

"404 - page not found The

master branch of

javafactory-plugin does not contain the path

docs/how-to-use.md."

How do I hook it into local models? Does it support Ollama, Continue, that kind of thing? Do you collect telemetry?