Skip to content
#

prompt-compression

Here are 37 public repositories matching this topic...

14-stage Fusion Pipeline for LLM token compression — reversible compression, AST-aware code analysis, intelligent content routing. Zero LLM inference cost. MIT licensed.

  • Updated Apr 1, 2026
  • Python

A curated list of strategies, tools, papers, and resources for reducing LLM token costs and improving efficiency in production.

  • Updated Apr 20, 2026

Enhance the performance and cost-efficiency of large-scale Retrieval Augmented Generation (RAG) applications. Learn to integrate vector search with traditional database operations and apply techniques like prefiltering, postfiltering, projection, and prompt compression.

  • Updated Jul 23, 2024
  • Jupyter Notebook

End-to-End Python implementation of CompactPrompt (Choi et al., 2025): a unified pipeline for LLM prompt and data compression. Features modular compression pipeline with dependency-driven phrase pruning, reversible n-gram encoding, K-means quantization, and embedding-based exemplar selection. Achieves 2-4x token reduction while preserving accuracy.

  • Updated Nov 30, 2025
  • Jupyter Notebook

PirateBao is a TypeScript/Bun agent-skill package for terse pirate-speak AI coding replies that preserve technical detail while cutting filler, with hooks, compressor CLI, OpenCode/Codex/Claude/Gemini cargo, .bao validation, npmjs gates, and token eval checks.

  • Updated Apr 13, 2026
  • TypeScript

Improve this page

Add a description, image, and links to the prompt-compression topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the prompt-compression topic, visit your repo's landing page and select "manage topics."

Learn more