Loading...
Loading...
Use when "deploying ML models", "MLOps", "model serving", "feature stores", "model monitoring", or asking about "PyTorch deployment", "TensorFlow production", "RAG systems", "LLM integration", "ML infrastructure"
npx skill4agent add eyadsibai/ltk ml-engineering| Category | Tools |
|---|---|
| ML Frameworks | PyTorch, TensorFlow, Scikit-learn, XGBoost |
| LLM Frameworks | LangChain, LlamaIndex, DSPy |
| Data Tools | Spark, Airflow, dbt, Kafka, Databricks |
| Deployment | Docker, Kubernetes, AWS/GCP/Azure |
| Monitoring | MLflow, Weights & Biases, Prometheus |
| Databases | PostgreSQL, BigQuery, Snowflake, Pinecone |
# Model serving with FastAPI
from fastapi import FastAPI
import torch
app = FastAPI()
model = torch.load("model.pth")
@app.post("/predict")
async def predict(data: dict):
tensor = preprocess(data)
with torch.no_grad():
prediction = model(tensor)
return {"prediction": prediction.tolist()}# Feast feature store
from feast import FeatureStore
store = FeatureStore(repo_path=".")
features = store.get_online_features(
features=["user_features:age", "user_features:location"],
entity_rows=[{"user_id": 123}]
).to_dict()# Drift detection
from evidently import ColumnMapping
from evidently.report import Report
from evidently.metric_preset import DataDriftPreset
report = Report(metrics=[DataDriftPreset()])
report.run(reference_data=ref_df, current_data=curr_df)| Metric | Target |
|---|---|
| P50 Latency | < 50ms |
| P95 Latency | < 100ms |
| P99 Latency | < 200ms |
| Throughput | > 1000 RPS |
| Availability | 99.9% |
# Basic RAG with LangChain
from langchain.vectorstores import Pinecone
from langchain.embeddings import OpenAIEmbeddings
from langchain.chains import RetrievalQA
vectorstore = Pinecone.from_existing_index(
index_name="docs",
embedding=OpenAIEmbeddings()
)
qa = RetrievalQA.from_chain_type(
llm=llm,
retriever=vectorstore.as_retriever()
)# Structured prompts with DSPy
import dspy
class QA(dspy.Signature):
"""Answer questions based on context."""
context = dspy.InputField()
question = dspy.InputField()
answer = dspy.OutputField()
qa = dspy.Predict(QA)# Development
python -m pytest tests/ -v --cov
python -m black src/
python -m pylint src/
# Training
python scripts/train.py --config prod.yaml
mlflow run . -P epochs=10
# Deployment
docker build -t model:v1 .
kubectl apply -f k8s/model-serving.yaml
# Monitoring
mlflow ui --port 5000