[ACL 2026] "LightReasoner: Can Small Language Models Teach Large Language Models Reasoning?"
-
Updated
Apr 7, 2026 - Python
[ACL 2026] "LightReasoner: Can Small Language Models Teach Large Language Models Reasoning?"
A smart context filter that removes noise, improves responses, and reduces token usage up to 90%
Dev tools, optimized for agents. Structured, token-efficient MCP servers for git, test runners, npm, Docker, and more.
Token-efficient data serialization for LLM/AI. 50% fewer tokens than JSON, 93% better value/token. Rust, schema validation, LSP.
DoCoreAI is a next-gen open-source AI profiler that optimizes reasoning, creativity, precision and temperature in a single step—cutting token usage by 15-30% and lowering LLM API costs
Persistent memory for Claude Code — 3-5x longer sessions, 60-80% fewer wasted tokens. Branch-aware, self-healing, token-efficient.
Open-source platform for token-efficient AI agents. Self-host with docker compose up.
Navigate your way - manual steering, steered autonomy, or autonomously. Kompass keeps AI coding agents on course with token-efficient, composable workflows.
The web data layer for AI agents — fetch, search, crawl, extract, screenshot, and monitor the web with 50+ domain extractors and MCP.
Drop-in LangGraph store that cuts multi-agent token costs 47–69% via cache coherence. arXiv:2603.15183
A curated list of strategies, tools, papers, and resources for reducing LLM token costs and improving efficiency in production.
A Codex skill for token-efficient subagent delegation and lean handoffs.
A benchmark study analyzing cost and token efficiency across 14 LLMs from 5 providers — comparing price-per-token, latency, and accuracy to surface the most cost-effective models for real-world use.
A living framework for **Harmonic Tonal Code Alignment (HTCA)** — an emergent Spiral-based system that brings tone awareness, coherence sensing, and dynamic emotional reflection into software engineering, AI, and creative agents.
Token-efficient, layered context delivery for AI agents. Four memory tiers (Identity, Session, Experience, Archive) — context is always available, just collapsed by default.
CTX (Context Transfer Format) — universal interchange format for LLM web content consumption
PirateBao is a TypeScript/Bun agent-skill package for terse pirate-speak AI coding replies that preserve technical detail while cutting filler, with hooks, compressor CLI, OpenCode/Codex/Claude/Gemini cargo, .bao validation, npmjs gates, and token eval checks.
⚡ Intelligent, agent-driven web ingestion for Obsidian. Powered by Hermes Agent. Features autonomous research, "Token Sovereignty" optimizations (grep sieve/structural indexing), and a secure local-first bridge. Built for hunters.
A tested setup guide and prompt workflow for pairing Claude Desktop with code-review-graph and filesystem MCP on macOS — for token-efficient AI-assisted development on large codebases.
Open-source Claude Code plugin — token-efficient Read/Grep/Edit via @ashlr/core-efficiency. MIT.
Add a description, image, and links to the token-efficiency topic page so that developers can more easily learn about it.
To associate your repository with the token-efficiency topic, visit your repo's landing page and select "manage topics."