Loading...
Loading...
Found 140 Skills
Guides PRD, ADR, Design Doc, and Work Plan creation with templates and decision matrix.
The Meta-Skill for managing standard capabilities. Use it to install, update, or list available standard skills (like auto-committer, context-code-explainer).
Privy SDK Documentation
Use when testing MCP servers, debugging MCP tool responses, exploring MCP capabilities, or diagnosing why an MCP tool returns unexpected data
Use this when you need to evaluate the risks and benefits of accepting, negotiating before accepting, pausing, or rejecting outsourcing projects, internal projects, or requirements. It is particularly suitable for scenarios with ambiguity in scope, acceptance criteria, payment terms, compliance, project timelines, or dependencies, as well as high-uncertainty situations such as emergency task insertion, contract renewal/modification, multi-requirement prioritization, or AI/LLM-related initiatives.
Design PM-friendly technical architecture for features. No code, only high-level design decisions.
Create detailed feature specifications with user stories, acceptance criteria, and edge cases. Use when starting a new feature or initializing a new project.
GENERator DNA 序列生成模型的昇腾 NPU 迁移 Skill,适用于将基于 HuggingFace Transformers 的 Causal LM 从 CUDA 迁移到华为 Ascend NPU,覆盖环境搭建、依赖安装、代码适配、多进程处理和 sequence recovery 验证。
Used for reviewing GitCode PRs, generating in-depth review conclusions or publishing line-by-line comments by combining PR metadata, diffs, and the context of the entire code repository. It is used when users want to review a GitCode PR, check a GitCode PR link, analyze change risks, or publish review comments to a GitCode PR. Typical trigger phrases include "review this PR", "inspect this PR", "check PR", or directly providing a GitCode PR link, such as https://gitcode.com/owner/repo/pull/123.
Calculate MFU (Machine FLOP Utilization) for operators such as matmul/GEMM, and provide clear formulas and derivation processes.
Migrate GPU/CUDA Triton operators to Triton-Ascend, or rewrite Python/PyTorch operators into Triton-Ascend implementations that can run on Ascend NPU. When clear optimization opportunities are identified, directly output the optimized code, minimal validation script, and troubleshooting instructions. This skill should be prioritized when users mention 昇腾 (Ascend), Ascend, NPU, triton-ascend, Triton operator migration, PyTorch operator rewriting, coreDim, UB overflow, 1D grid, physical core binding, block_ptr, stride, memory access alignment, mask performance, dtype degradation, operator optimization, or directly ask questions like "How to use this skill", "How to run it in the command line", "How to perform migration/validation in a container", even if users do not explicitly say "write a skill" or "perform migration".
AI for Science 场景下的昇腾 NPU Profiling 采集与性能分析 Skill,用于在华为 Ascend NPU 上使用 torch_npu.profiler 采集 L0、L1、L2 级性能数据,分析训练或推理中的算子耗时、调用栈、内存与瓶颈,并指导后续调优。