Loading...
Loading...
Found 5 Skills
Build real-time conversational AI voice engines using async worker pipelines, streaming transcription, LLM agents, and TTS synthesis with interrupt handling and multi-provider support
LangGraph tool calling patterns. Use when binding tools to LLMs, implementing ToolNode for execution, dynamic tool selection, or adding approval gates to tool calls.
Inter-agent communication patterns including message passing, shared memory, blackboard systems, and event-driven architectures for LLM agentsUse when "agent communication, message passing, inter-agent, blackboard, agent events, multi-agent, communication, message-passing, events, coordination" mentioned.
Consult an advisory council of three AI personas — Cato (skeptic), Ada (optimist), Marcus (pragmatist) — backed by different frontier LLM agents (Gemini, Claude, Codex). Each persona runs as a separate agent process with full repo context and returns independent feedback. Use when the user says "/council", asks for a second opinion, wants feedback on code changes, needs a premortem, wants to pressure-test a decision, or asks "what do you think about this approach?" Claude may also proactively suggest consulting the council before major architectural decisions, risky deploys, or ambiguous trade-offs (but should ask for user approval first).
Apply Model-First Reasoning (MFR) to code generation tasks. Use when the user requests "model-first", "MFR", "formal modeling before coding", "model then implement", or when tasks involve complex logic, state machines, constraint systems, or any implementation requiring formal correctness guarantees. Enforces strict separation between modeling and implementation phases.