Awesome-LLM-Prompt-Optimization: a curated list of advanced prompt optimization and tuning methods in Large Language Models
-
Updated
Mar 27, 2024
Awesome-LLM-Prompt-Optimization: a curated list of advanced prompt optimization and tuning methods in Large Language Models
Superpipe - optimized LLM pipelines for structured data
Implementation of the paper Fast Inference from Transformers via Speculative Decoding, Leviathan et al. 2023.
Nadir is a Python package designed to dynamically choose the best llm for your prompt by balancing complexity and cost and response time.
A Demo of Running Sleep-time Compute to Reduce LLM Latency
DSPEx - Declarative Self-improving Elixir | A BEAM-Native AI Program Optimization Framework
Declarative Self Improving Elixir - DSPy Orchestration in Elixir
Generate clean, AI-ready llms.txt files for your website or docs. Supports crawling, sitemaps, static builds, and framework-aware adapters (Next.js, Vite, Nuxt, Astro, Remix). Includes Markdown/MDX docs mode and robots.txt generator for LLM and search crawlers.
⚡ Cut LLM inference costs 80% with Programmatic Tool Calling. Instead of N tool call round-trips, generate JavaScript to orchestrate tools in Vercel Sandbox. Supports Anthropic, OpenAI, 100+ models via AI Gateway. Novel MCP Bridge for external service integration.
Solve the MCP token bloat problem. Build lean, custom MCP servers with only the tools you need from 300+ available tools. Reduce LLM costs and hallucinations. Includes CLI, WebSocket proxy, AI-powered recommendations, and interactive playground.
🚄 Advanced LLM optimization techniques using CUDA. Features efficient attention mechanisms, custom CUDA kernels for transformers, and memory-efficient training strategies. ⚡
Research into neural network quantization with focus on domain-specific calibration and hardware-aware optimization. Committed to rigorous methodology and reproducible results.
An intelligent prompt optimizer that tailors your prompts for specific LLMs like Gemini, Claude (Anthropic), and ChatGPT using advanced prompt engineering techniques.
LLM-optimised prompt library for BusinessMate — the AI strategy coach for small business owners (SMB's) and solopreneurs.
Track your website’s LLM ranking performance with real-time insights.
End-to-End Python implementation of CompactPrompt (Choi et al., 2025): a unified pipeline for LLM prompt and data compression. Features modular compression pipeline with dependency-driven phrase pruning, reversible n-gram encoding, K-means quantization, and embedding-based exemplar selection. Achieves 2-4x token reduction while preserving accuracy.
DocuMerge for LLMs 🚀 Easily prepare GitHub repositories for LLM analysis!
Zero-config Cloudflare Worker for LunaLift integration. Serves llms.txt, injects schemas, tracks AI/LLM traffic. Requires LunaLift subscription.
A REST API server for DSPy that enables language model programming from any language. This proxy exposes DSPy's core functionality - signature registration, module execution, optimization, and evaluation - through HTTP endpoints, making it accessible from non-Python environments.
The Global Open Knowledge Graph for AI & LLMs. A verified registry of structured data (JSON-LD) providing canonical entity resolution and digital identity for organizations, events, and experts. CC0 License.
Add a description, image, and links to the llm-optimization topic page so that developers can more easily learn about it.
To associate your repository with the llm-optimization topic, visit your repo's landing page and select "manage topics."