Loading...
Loading...
Found 7 Skills
Interact with KWeaver Knowledge Network and Decision Agent — build knowledge networks, query Schema/instances, semantic search, execute Action, Agent CRUD and conversation, Trace data analysis. Interact with Dataflow document processes — list processes, trigger runs, query run history, view step logs. Interact with Skill management module — register Skill, search in market, progressive reading, download and installation. Interact with Toolbox / Tool — create toolbox, upload OpenAPI tools, publish, start and stop. Interact with Vega observability platform — query Catalog/resources/connector types, health inspection. This skill is automatically activated when users mention intents such as "knowledge network", "knowledge graph", "query object type", "execute Action", "what Agents are there", "create Agent", "converse with Agent", "list all Agent templates", "list Agents I created", "list Agents in private space", "dataflow", "data flow", "process orchestration", "process run records", "process logs", "trigger dataflow", "view dataflow run history", "Skill", "skill package", "register Skill", "install Skill", "read SKILL.md", "toolbox", "toolbox", "upload tool", "register tool", "OpenAPI tool", "enable tool", "publish toolbox", "data source", "data view", "atomic view", "Catalog", "Vega", "health check", "inspection", "trace", "evidence chain", "data flow tracking", "data source", "how data is obtained", etc.
Primary entry point for building, managing, and orchestrating data pipelines on Google Cloud. Guides users to the appropriate skill for dbt, Dataflow (Apache Beam), Dataform, Spark (Dataproc Serverless), BigQuery Data Transfer Service (DTS) or orchestration pipeline using Cloud Composer. Clarify requirements and resolve ambiguity for creating, updating and running data pipelines.
Provides guidance for writing, packaging and executing Apache Beam pipelines on GCP using Cloud Dataflow. Use when: - Creating an Apache Beam Dataflow pipeline. - Creating a Google Flex Template.
Google Cloud Dataflow integration. Manage data, records, and automate workflows. Use when the user wants to interact with Google Cloud Dataflow data.
Guided, interactive exploration of statistical data via SDMX providers (Eurostat, OECD, ECB, World Bank, ISTAT, and others) using the opensdmx CLI. Use this skill whenever the user asks ANY question about statistics or data that could be answered with SDMX data — even if they don't mention SDMX, Eurostat, or any provider by name. Topics include demographics, economy, employment, births, deaths, population, prices, trade, health, agriculture, GDP, inflation, unemployment, fertility rates, migration, energy, education, poverty, housing, and any other statistical topic. Also use it when the user mentions a specific dataflow ID they want to explore. Trigger this skill even for implicit questions like "how many births were there in Italy last year?" or "I need EU unemployment data by age group" — these clearly need SDMX data even if the user doesn't say so. The skill guides the user step by step: discovers relevant datasets, proposes the most meaningful candidates, explores the schema using real constraints (not codelists), explains the dataset structure, and invites the user to make informed filter choices before fetching any data.
Design, refactor, and review Effector state management using modern v23+ patterns. Use when tasks involve createStore/createEvent/createEffect modeling, dataflow with sample/attach/split, scope-safe SSR with fork/allSettled/serialize/hydrate, React integration with useUnit, Solid/Vue integration patterns, fixing scope loss, or replacing anti-patterns such as business logic in watch, imperative calls in effects, and direct getState business reads.
Guides understanding and working with Apache Beam runners (Direct, Dataflow, Flink, Spark, etc.). Use when configuring pipelines for different execution environments or debugging runner-specific issues.