180 results for “topic:token-optimization”
CLI proxy that reduces LLM token consumption by 60-90% on common dev commands. Single Rust binary, zero dependencies
🦞 LLM Token Compression & Reduction Tool — Cut AI agent token costs by up to 97%. 5-layer context compression for OpenClaw workspaces. Deterministic, no LLM required. Token saving, prompt compression, context window optimization.
The Context Optimization Layer for LLM Applications
Working memory for Claude Code - persistent context and multi-instance coordination
An MCP server that executes Python code in isolated rootless containers with optional MCP server proxying. Implementation of Anthropic's and Cloudflare's ideas for reducing MCP tool definitions context bloat.
Production-ready modular Claude Code framework with 30+ commands, token optimization, and MCP server integration. Achieves 2-10x productivity gains through systematic command organization and hierarchical configuration.
TOON encoding for Laravel. Encode data for AI/LLMs with ~50% fewer tokens than JSON.
Config-driven CLI tool that compresses command output before it reaches an LLM context
TOON — Laravel AI package for compact, human-readable, token-efficient data format with JSON ⇄ TOON conversion for ChatGPT, OpenAI, and other LLM prompts.
Find the ghost tokens. Audit your Claude Code context window overhead, see where tokens go, get them back.
Multi-agent orchestration for Claude Code with 15-30% token optimization, self-improving agents, and automatic verification
No description provided.
RustAPI – A batteries-included Rust web framework with FastAPI-like ergonomics, OpenAPI docs, JWT, and MCP-ready TOON format for AI & LLM APIs.
OCTAVE protocol - structured AI communication with 3-20x token reduction. MCP server with lenient-to-canonical pipeline and schema validation.
🚀 Lightweight Python library for building production LLM applications with smart context management and automatic token optimization. Save 10-20% on API costs while fitting RAG docs, chat history, and prompts into your token budget.
Context Limiter & Output Vetter for context bloat. It is a highly specialized, structure-aware JSON built specifically to intercept and compress MCP responses before they annihilate your LLM's context window.
Claude Code plugin that offloads large outputs to filesystem and retrieves when required.
Security hooks and monitoring for Claude Code — quiet overrides, SSRF protection, MCP compression, OTEL tracing
Laravel integration for TOON format: encode/decode JSON data into a token-optimized notation for LLMs.
Intelligent token optimization for Claude Code - achieving 95%+ token reduction through caching, compression, and smart tool intelligence
💰 Save money on AI API costs! 76% token reduction, Auto-Fix token limits, Universal AI compatibility. Cline • Copilot • Claude • Cursor
🏗️ AI-friendly Node.js project architecture standards. Keep files <400 lines for AI agents. Covers H5 games, data tools, APIs, SDKs. 70-93% token savings. OpenClaw skill.
⚡ Cut LLM inference costs 80% with Programmatic Tool Calling. Instead of N tool call round-trips, generate JavaScript to orchestrate tools in Vercel Sandbox. Supports Anthropic, OpenAI, 100+ models via AI Gateway. Novel MCP Bridge for external service integration.
⚜️ An MCP server for context compaction and recycling in Claude Code
🦞 龙虾饼 - 智能上下文压缩系统,让 AI 记忆永不溢出
CLI proxy that cuts noisy terminal output while preserving command behavior
Token Oriented Object Notation (TOON) for Linked Data
Optimized file search and semantic indexing for large codebases in Claude Code
A declarative build system for complex multi-step autonomous builds
OpenClaw skill for offloading heavy tasks to NVIDIA NIM models and saving context tokens