Awesome-LLM-Prompt-Optimization: a curated list of advanced prompt optimization and tuning methods in Large Language Models
-
Updated
Mar 27, 2024
Awesome-LLM-Prompt-Optimization: a curated list of advanced prompt optimization and tuning methods in Large Language Models
Superpipe - optimized LLM pipelines for structured data
Implementation of the paper Fast Inference from Transformers via Speculative Decoding, Leviathan et al. 2023.
Nadir is a Python package designed to dynamically choose the best llm for your prompt by balancing complexity and cost and response time.
Optimize AI workflows with Arachne. Automatically assembles the perfect code context (Tree, Target, Deps, Semantic) to fit context windows without noise. Built for efficiency and scale.
Optimized fork of Superpowers — Retains full original features while adding automatic 3-tier workflow routing, integrated safety guards (OWASP-aligned), red-team adversarial testing with auto-fix pipeline, a built-in memory stack, and much more. Delivers faster, more reliable, hallucination-resistant coding sessions.
Credit Optimizer v5 for Manus AI - Save 30-75% on credits. Free MCP Server (PyPI) + $9 Manus Skill. Audited across 53 scenarios. Zero quality loss.
DSPEx - Declarative Self-improving Elixir | A BEAM-Native AI Program Optimization Framework
A Demo of Running Sleep-time Compute to Reduce LLM Latency
Declarative Self Improving Elixir - DSPy Orchestration in Elixir
⚡ Cut LLM inference costs 80% with Programmatic Tool Calling. Instead of N tool call round-trips, generate JavaScript to orchestrate tools in Vercel Sandbox. Supports Anthropic, OpenAI, 100+ models via AI Gateway. Novel MCP Bridge for external service integration.
Generate clean, AI-ready llms.txt files for your website or docs. Supports crawling, sitemaps, static builds, and framework-aware adapters (Next.js, Vite, Nuxt, Astro, Remix). Includes Markdown/MDX docs mode and robots.txt generator for LLM and search crawlers.
Hybrid adaptive memory system for Claude Code — Cheatsheet (positive patterns) + Immune (negative patterns)
A comprehensive toolkit for training and running lightweight adapters for GGUF-based language models (ERNIE, Llama, Mistral, Phi-3, etc.) without modifying the base model.
Opti-Oignon is a comprehensive optimization framework for local LLMs running on Ollama. It maximizes the performance of your local models through intelligent task routing based on a custom benchmark, RAG (Retrieval-Augmented Generation), and multi-model orchestration.
End-to-End Python implementation of CompactPrompt (Choi et al., 2025): a unified pipeline for LLM prompt and data compression. Features modular compression pipeline with dependency-driven phrase pruning, reversible n-gram encoding, K-means quantization, and embedding-based exemplar selection. Achieves 2-4x token reduction while preserving accuracy.
Adaptive semantic cache for LLMs with streaming support, ML-based thresholds, and real-time cost tracking. Built in Rust for sub-millisecond performance.
Structured repository covering LLM foundations, fine-tuning workflows, optimization strategies, deployment patterns, evaluation methods, and Responsible AI considerations.
Bio-inspired optimization pipeline for Claude Code. 3 systems: Slime Mold (explore→prune) → PRISM (perspectives→compile) → Immune (scan→correct→learn). Domain-agnostic.
LLM context compression proxy — 40-70% token savings, zero code changes
Add a description, image, and links to the llm-optimization topic page so that developers can more easily learn about it.
To associate your repository with the llm-optimization topic, visit your repo's landing page and select "manage topics."