Loading...
Loading...
Found 2 Skills
Optimizes text, prompts, and documentation for LLM token efficiency. Applies 41 research-backed rules across 6 categories: Claude behavior, token efficiency, structure, reference integrity, perception, and LLM comprehension. Use when optimizing prompts, reducing tokens, compressing verbose docs, or improving LLM instruction quality.
Learn how to manage conversation context in AMCP to avoid LLM API errors from exceeding context windows. This skill covers SmartCompactor strategies, token estimation, configuration, and best practices.