Loading...
Loading...
Found 30 Skills
Guide for making code reviews. Use this when asked to make code reviews, or ask to use it before committing changes.
Execute deep web research on any topic using Yutori's research agents. Use for competitive analysis, market research, finding documentation, or answering complex questions that require synthesizing information from multiple sources.
Scan skills to extract cross-cutting principles and distill them into rules — append, revise, or create new rule files
Use when reviewing React documentation for structure, components, and style compliance
Research-aligned self-consistency for debugging. Spawns independent solver agents that each explore and debug the problem from scratch. Uses majority voting. Based on "Self-Consistency Improves Chain of Thought Reasoning" (Wang et al., 2022). Use for critical bugs, algorithms, or when other approaches have failed.
Comprehensive review of local uncommitted changes using specialized agents with code improvement suggestions
Generate a "Journey Into [Project]" narrative report analyzing a project's entire development history from claude-mem's timeline. Use when asked for a timeline report, project history analysis, development journey, or full project report.
Nuclear-grade 16-agent pre-publish release gate. Runs /get-unpublished-changes to detect all changes since last npm release, spawns up to 10 ultrabrain agents for deep per-change analysis, invokes /review-work (5 agents) for holistic review, and 1 oracle for overall release synthesis. Use before EVERY npm publish. Triggers: 'pre-publish review', 'review before publish', 'release review', 'pre-release review', 'ready to publish?', 'can I publish?', 'pre-publish', 'safe to publish', 'publishing review', 'pre-publish check'.
Intelligent multi-topic in-depth research tool that supports input of any materials, uses independent research Agents for parallel in-depth retrieval and generates systematic research documents. This skill should be used when users need to conduct in-depth research on multiple related topics, perform systematic information retrieval, and integrate multi-angle analysis.
MantaBase T3 Hardware Audit System (Chinese Version). Objectively classify hardware products based on design theories through Brand Blinding, three Auditor (Tool/Toy/Trash) special scoring and Peer Review. Triggers: product link, T3 audit, Tool/Toy/Trash classification, hardware evaluation, VC investment advice
Constructive critique through 5 HackerNews commenter personas with evidence-based claim validation. Use when user wants devil's advocacy, stress testing, or critical review of ideas, docs, architecture, or code. Use for "roast", "critique this", "poke holes", "devil's advocate", "stress test", or "what's wrong with". Do NOT use for code review (use systematic-code-review), implementation changes, or performance profiling without a specific critique request.
Adversarial Quarto vs Beamer QA. Critic finds issues, fixer applies fixes, loops until APPROVED (max 5 rounds).