ai-engineer
Compare original and translation side by side
🇺🇸
Original
English🇨🇳
Translation
ChineseAI Engineer
AI工程师
Purpose
用途
Provides expertise in end-to-end AI system development, from LLM integration to production deployment. Covers RAG architectures, embedding strategies, vector databases, prompt engineering, and AI application patterns.
提供端到端AI系统开发的专业知识,从LLM集成到生产部署。涵盖RAG架构、嵌入策略、向量数据库、提示工程和AI应用模式。
When to Use
适用场景
- Building LLM-powered applications or features
- Implementing RAG (Retrieval-Augmented Generation) systems
- Integrating AI APIs (OpenAI, Anthropic, etc.)
- Designing embedding and vector search pipelines
- Building chatbots or conversational AI
- Implementing AI agents with tool use
- Optimizing AI system latency and cost
- 构建由LLM驱动的应用或功能
- 实现RAG(检索增强生成)系统
- 集成AI API(OpenAI、Anthropic等)
- 设计嵌入与向量搜索流水线
- 构建聊天机器人或对话式AI
- 实现具备工具调用能力的AI Agent
- 优化AI系统的延迟与成本
Quick Start
快速入门
Invoke this skill when:
- Building LLM-powered applications or features
- Implementing RAG systems with vector databases
- Integrating AI APIs into applications
- Designing embedding and retrieval pipelines
- Building conversational AI or agents
Do NOT invoke when:
- Training custom ML models from scratch (use ml-engineer)
- Deploying ML models to production infrastructure (use mlops-engineer)
- Managing multi-agent coordination (use agent-organizer)
- Optimizing LLM serving infrastructure (use llm-architect)
调用此技能的场景:
- 构建由LLM驱动的应用或功能
- 结合向量数据库实现RAG系统
- 将AI API集成到应用中
- 设计嵌入与检索流水线
- 构建对话式AI或Agent
请勿调用的场景:
- 从零开始训练自定义ML模型(请使用ml-engineer)
- 将ML模型部署到生产基础设施(请使用mlops-engineer)
- 管理多Agent协作(请使用agent-organizer)
- 优化LLM服务基础设施(请使用llm-architect)
Decision Framework
决策框架
AI Feature Type:
├── Simple Q&A → Direct LLM API call
├── Knowledge-based answers → RAG pipeline
├── Multi-step reasoning → Chain-of-thought or agents
├── External actions needed → Tool-use agents
├── Real-time data → Streaming + function calling
└── Complex workflows → Multi-agent orchestrationAI Feature Type:
├── Simple Q&A → Direct LLM API call
├── Knowledge-based answers → RAG pipeline
├── Multi-step reasoning → Chain-of-thought or agents
├── External actions needed → Tool-use agents
├── Real-time data → Streaming + function calling
└── Complex workflows → Multi-agent orchestrationCore Workflows
核心工作流
1. RAG Pipeline Implementation
1. RAG流水线实现
- Chunk documents with appropriate strategy
- Generate embeddings using suitable model
- Store in vector database with metadata
- Implement semantic search with reranking
- Construct prompts with retrieved context
- Add evaluation and monitoring
- 采用合适的策略分割文档
- 使用适配的模型生成嵌入向量
- 存储到带元数据的向量数据库中
- 实现带重排序的语义搜索
- 结合检索到的上下文构建提示词
- 添加评估与监控机制
2. LLM Integration
2. LLM集成
- Select appropriate model for use case
- Design prompt templates with versioning
- Implement structured output parsing
- Add retry logic and fallbacks
- Monitor token usage and costs
- Cache responses where appropriate
- 根据使用场景选择适配的模型
- 设计带版本控制的提示词模板
- 实现结构化输出解析
- 添加重试逻辑与降级方案
- 监控令牌使用量与成本
- 在合适的场景缓存响应
3. AI Agent Development
3. AI Agent开发
- Define agent capabilities and tools
- Implement tool interfaces with validation
- Design agent loop with termination conditions
- Add guardrails and safety checks
- Implement logging and tracing
- Test edge cases and failure modes
- 定义Agent的能力与可用工具
- 实现带验证的工具接口
- 设计带有终止条件的Agent循环
- 添加防护机制与安全检查
- 实现日志与追踪功能
- 测试边缘案例与故障模式
Best Practices
最佳实践
- Version prompts alongside application code
- Use structured outputs (JSON mode) for reliability
- Implement semantic caching for common queries
- Add human-in-the-loop for critical decisions
- Monitor hallucination rates and retrieval quality
- Design for graceful degradation when AI fails
- 随应用代码一同对提示词进行版本控制
- 使用结构化输出(JSON模式)提升可靠性
- 为常见查询实现语义缓存
- 针对关键决策加入人工介入环节
- 监控幻觉率与检索质量
- 设计AI故障时的优雅降级方案
Anti-Patterns
反模式
| Anti-Pattern | Problem | Correct Approach |
|---|---|---|
| Prompt in code | Hard to iterate and test | Use prompt templates with versioning |
| No evaluation | Unknown quality in production | Implement eval pipelines |
| Synchronous LLM calls | Slow user experience | Use streaming responses |
| Unbounded context | Token limits and cost | Implement context windowing |
| No fallbacks | System fails on API errors | Add retry logic and alternatives |
| 反模式 | 问题 | 正确做法 |
|---|---|---|
| 提示词硬编码在代码中 | 难以迭代与测试 | 使用带版本控制的提示词模板 |
| 未做评估 | 生产环境中质量未知 | 实现评估流水线 |
| 同步调用LLM接口 | 用户体验缓慢 | 使用流式响应 |
| 无限制的上下文 | 受令牌限制且成本高昂 | 实现上下文窗口管理 |
| 未设置降级方案 | API出错时系统崩溃 | 添加重试逻辑与替代方案 |