Loading...
Loading...
Found 21 Skills
Query the ExoPriors Scry API -- SQL-over-HTTPS search across 229M+ entities spanning forums, papers, social media, government records, and prediction markets. Includes cross-platform author identity resolution (actors, people, aliases), OpenAlex academic graph navigation (authors, citations, institutions, concepts), shareable artifacts, and structured agent judgements. Use when the task involves: Scry API, ExoPriors, /v1/scry/query, scry.search, scry.entities, materialized views, corpus search, epistemic infrastructure, 229M entities, lexical search, BM25, structured agent judgements, scry shares, cross-corpus analysis, who is this person, cross-platform identity, OpenAlex, citation graph, coauthor graph, academic papers, author lookup. NOT for: semantic/vector search composition or embedding algebra (use scry-vectors), LLM-based reranking (use scry-rerank), or the user's own local Postgres / non-ExoPriors data sources.
Use this skill when designing data warehouses, building star or snowflake schemas, implementing slowly changing dimensions (SCDs), writing analytical SQL for Snowflake or BigQuery, creating fact and dimension tables, or planning ETL/ELT pipelines for analytics. Triggers on dimensional modeling, surrogate keys, conformed dimensions, warehouse architecture, data vault, partitioning strategies, materialized views, and any task requiring OLAP schema design or warehouse query optimization.
Use this skill when building real-time or near-real-time data pipelines. Covers Kafka, Flink, Spark Streaming, Snowpipe, BigQuery streaming, materialized views, and batch-vs-streaming decisions. Common phrases: "real-time pipeline", "Kafka consumer", "streaming vs batch", "low latency ingestion". Do NOT use for batch integration patterns (use integration-patterns-skill) or pipeline orchestration (use data-orchestration-skill).
Configure data accelerators for local materialization and caching in Spice (Arrow, DuckDB, SQLite, Cayenne, PostgreSQL, Turso). Use when asked to "accelerate data", "enable caching", "materialize dataset", "configure refresh", "set up local storage", "improve query performance", "choose an accelerator", or "configure snapshots".
Author ZenML pipelines: @step/@pipeline decorators, type hints, multi-output steps, dynamic vs static pipelines, artifact data flow, ExternalArtifact, YAML configuration, DockerSettings for remote execution, custom materializers, metadata logging, secrets management, and custom visualizations. Use this skill whenever asked to write a ZenML pipeline, create ZenML steps, make a pipeline work on Kubernetes/Vertex/SageMaker, add Docker settings, write a materializer, create a custom visualization, handle "works locally but fails on cloud" issues, or configure pipeline YAML files. Even if the user doesn't explicitly mention "pipeline authoring", use this skill when they ask to build an ML workflow, data pipeline, or training pipeline with ZenML.
Use this skill whenever working with QuestDB — a high-performance time-series database. Trigger on any mention of QuestDB, time-series SQL with SAMPLE BY, LATEST ON, ASOF JOIN, ILP ingestion, or the questdb Python/Go/Java/Rust/.NET client libraries. Also trigger when writing Grafana queries against QuestDB, creating materialized views for time-series rollups, working with order book or financial market data in QuestDB, or any SQL that involves designated timestamps or time-partitioned tables. QuestDB extends SQL with unique time-series keywords — standard PostgreSQL or MySQL patterns will fail. Always read this skill before writing QuestDB SQL to avoid hallucinating incorrect syntax.
129 practical Oracle Database and Oracle Container Registry reference guides covering SQL/PL/SQL development, performance tuning (AWR, ASH, explain plan, indexes, wait events, memory), security (TDE, VPD, auditing, network), administration (RMAN, Data Guard, undo/redo, users), monitoring, architecture (RAC, CDB/PDB, Exadata, In-Memory, OCI), DevOps (Liquibase, Flyway, utPLSQL, EBR), migrations from Postgres/MySQL/SQL Server/MongoDB/Snowflake/Redshift/DB2, PL/SQL development (packages, cursors, collections, unit testing, debugging), Oracle features (AQ, DBMS_SCHEDULER, materialized views, APEX), SQLcl (basics, scripting, Liquibase, MCP server, CI/CD), ORDS (architecture, authentication, AutoREST, REST API design, PL/SQL gateway), and Oracle Container Registry images. Use for any Oracle DB question, ORA- errors, DBMS_ packages, v$ views, Oracle tooling, ORDS REST APIs, SQLcl commands, or Oracle container images. Always consult this skill before answering Oracle-specific questions.
Diagnose ClickHouse Kafka engine health, consumer status, thread pool capacity, and consumption issues. Use for Kafka lag, consumer errors, and thread starvation.
Diagnose ClickHouse INSERT performance, batch sizing, part creation patterns, and ingestion bottlenecks. Use for slow inserts and data pipeline issues.