Loading...
Loading...
Found 3,563 Skills
Integrate multiple plot point analysis results into a comprehensive report, and generate high-quality analysis through deduplication, classification, sorting, and summarization. Suitable for integrating multiple analysis sources and generating unified reports
Создание текстовых документов. Загрузи этот skill ПЕРЕД созданием документа, статьи, отчёта или письма.
Analyze multi-round evaluation score data, count various indicators, and calculate rating levels. Suitable for analyzing score trends and calculating S/A/B ratings
Analyze story genre types, extract creative elements and story features. Suitable for identifying genre positioning, refining creativity, and analyzing market audiences
Skill for PHP/Laravel backend development following project conventions. Use when creating or editing PHP code, models, services, controllers, tests, or any backend logic. Loads all backend rules from .claude/rules/backend/ and .claude/rules/dataclasses/.
Skill for writing and updating codebase documentation. Use when creating or editing markdown documentation files in the docs/ directory, README files, or any documentation-related content. Also activates when maintaining the documentation index.
Skill for creating and editing PHP tests following project conventions. Use when creating tests, updating test files, or refactoring tests. Applies proper structure, naming, factory usage, and Laravel/PHPUnit best practices.
Observability and monitoring for data pipelines using OpenTelemetry (traces) and Prometheus (metrics). Covers instrumentation, dashboards, and alerting.
Using DuckDB with remote cloud storage via HTTPFS extension, fsspec, and Delta Lake integration. Covers S3, GCS, Azure, and S3-compatible endpoints.
Native Arrow filesystem integration with PyArrow. Optimized for Parquet workflows, zero-copy data transfer, predicate pushdown, and column pruning. Covers S3, GCS, HDFS with PyArrow datasets.
Interactive web apps for data science: Streamlit, Panel, and Gradio. Use for prototyping ML models, creating data exploration dashboards, and sharing insights with non-technical stakeholders.
Delta Lake integration with cloud storage (S3, GCS, Azure). Covers storage_options, PyArrow filesystem, time travel, and partitioned writes.