Show HN: A deterministic middleware to compress LLM prompts by 50-80%

via github.com

Short excerpt below. Read at the original source.

Hi HN, I’m working on Skillware, an open-source framework that treats AI capabilities as installable, self-contained modules. I just added a “Prompt Token Rewriter” skill. It’s an offline heuristic middleware that strips conversational filler and redundant context from long agentic loops before they hit the LLM. It saves significant token costs and inference time, and […]

Read at Source