Show HN: Compliant LLM toolkit for ensuring compliance & security of AI systems

7 kaushik92 0 6/3/2025, 2:51:51 PM github.com ↗
With the right technique, I was able to break the so-called secure models like Claude and OpenAI.

So, I built an open-source tool to automate this and find security holes in any hosted model.

I got claude-sonnet-4 to demonstrate the following harmful behavior:

- steal data from downstream tool calls using sql injection, code injection and template injection attacks

- install spyware or malware using prompt obfuscation to send data to a third-party server

Try it yourself with this simple command:

  pip install compliant-llm && compliant-llm dashboard

Comments (0)

No comments yet