[ACL 2026] "LightReasoner: Can Small Language Models Teach Large Language Models Reasoning?"
-
Updated
Apr 7, 2026 - Python
[ACL 2026] "LightReasoner: Can Small Language Models Teach Large Language Models Reasoning?"
Dev tools, optimized for agents. Structured, token-efficient MCP servers for git, test runners, npm, Docker, and more.
Token-efficient data serialization for LLM/AI. 50% fewer tokens than JSON, 93% better value/token. Rust, schema validation, LSP.
A smart context filter that removes noise, improves responses, and reduces token usage up to 90%
DoCoreAI is a next-gen open-source AI profiler that optimizes reasoning, creativity, precision and temperature in a single step—cutting token usage by 15-30% and lowering LLM API costs
Persistent memory for Claude Code — 3-5x longer sessions, 60-80% fewer wasted tokens. Branch-aware, self-healing, token-efficient.
Open-source platform for token-efficient AI agents. Self-host with docker compose up.
Navigate your way - manual steering, steered autonomy, or autonomously. Kompass keeps AI coding agents on course with token-efficient, composable workflows.
The web data layer for AI agents — fetch, search, crawl, extract, screenshot, and monitor the web with 50+ domain extractors and MCP.
A Codex skill for token-efficient subagent delegation and lean handoffs.
A benchmark study analyzing cost and token efficiency across 14 LLMs from 5 providers — comparing price-per-token, latency, and accuracy to surface the most cost-effective models for real-world use.
A living framework for **Harmonic Tonal Code Alignment (HTCA)** — an emergent Spiral-based system that brings tone awareness, coherence sensing, and dynamic emotional reflection into software engineering, AI, and creative agents.
Token-efficient, layered context delivery for AI agents. Four memory tiers (Identity, Session, Experience, Archive) — context is always available, just collapsed by default.
PirateBao is a TypeScript/Bun agent-skill package for terse pirate-speak AI coding replies that preserve technical detail while cutting filler, with hooks, compressor CLI, OpenCode/Codex/Claude/Gemini cargo, .bao validation, npmjs gates, and token eval checks.
The Semantic Turning Point Detector is a lightweight but powerful tool for detecting semantic turning points in conversations or textual sequences. It recursively analyzes message chains (dialogues, transcripts, chat logs) and identifies where key shifts in meaning, topic, or insight occur.
CTX (Context Transfer Format) — universal interchange format for LLM web content consumption
Efficient web information retrieval and summation without excessive token usage.
Agent-to-Agent coordination that doesn't waste your context window. Token-efficient protocol with progressive discovery, zero-schema invocation, gRPC transport, and task lifecycle management.
Convert JSON format to TOON
This living repo documents academic exploration of AI architecture, token efficiency, and prompt engineering best practices.
Add a description, image, and links to the token-efficiency topic page so that developers can more easily learn about it.
To associate your repository with the token-efficiency topic, visit your repo's landing page and select "manage topics."