Show HN: Clean Clode – Clean Messy Terminal Pastes from Claude Code and Codex

4 thewojo 2 9/16/2025, 4:30:53 PM cleanclode.com ↗
I’ve been impressed with Claude Code but one thing that sometimes gets in the way in my workflows is the messy, mangled text that is shown when pasting text from the Claude Code terminal sessions. So I built an open-source utility that cleans extraneous white space, pipes, and other characters from your CC/Codex pastes.

For example, you can turn this:

`How can I create a Claude Code script that │ │ cleans up extraneous characters and cleans up │ │ extra spaces, new lines, and other messiness │

when I copy from Claude Code terminal │ │ prompts or copy code from Claude Responses in │ │ the Claude Code Terminal? It can make it │ │ hard to read, save, and reuse. `

Into this:

`How can I create a Claude Code script that cleans up extraneous characters and cleans up extra spaces, new lines, and other messiness when I copy from Claude Code terminal prompts or copy code from Claude Responses in the Claude Code Terminal? It can make it hard to read, save, and reuse. While this was built with Claude Code in mind it also works on Codex.`

Try it here: https://cleanclode.com

It’s 100% private (no data collection, tracking, completely open-source). If there’s anything you don’t like please just create a GitHub issue, contribute your change (https://github.com/TheJoWo/Clean-Clode), or comment here. Thanks and hope it’s helpful to some of you

Comments (2)

westurner · 37m ago
From https://github.com/google-gemini/gemini-cli/pull/5342#issuec... :

> Would .ipynb format solve for this? Unfortunately there's not yet a markdown format that includes output cells (likely due to the unusability of base64 encoded binary data). There are existing issues TODO to create a new format for Jupyter notebooks; which have notebook-level metadata, cell-level metadata, input cells, and output cells.

API facades like OpenLLM and model routers like OpenRouter have standard interfaces for many or most LLM inputs and outputs. Tools like Promptfoo, ChainForge, and LocalAI also all have abstractions over many models.

What are the open standards for representing LLM inputs, and outputs?

W3C PROV has prov:Entity, prov:Activity, and prov:Agent for modeling AI provenance: who or what did what when.

LLM evals could be represented in W3C EARL Evaluation and Reporting Language.

From https://news.ycombinator.com/item?id=44934531 :

> simonw/llm by default saves all prompt inputs and outputs in a sqlite database. Copilot has /save and gemini-cli has /export, but they don't yet autosave or flush before attempting to modify code given the prompt output?*

westurner · 1h ago
Here's a script to parse inputs out of Gemini CLI saved chats from before they implemented the /export command:

parsegeminiclisaves.sh: https://github.com/westurner/dotfiles/blob/933e89e7664e58225...