AI Agentic Marketing Workflows
<!-- dual-compat:start -->
Use when
- Design and implement autonomous AI marketing agent systems using the PRAL, BDI, and OODA frameworks. Invoke when a client is ready to move beyond reactive GenAI prompting to proactive, autonomous marketing workflows, or when planning an AI-first marketing operations architecture.
- Use this skill when it is the closest match to the requested deliverable or workflow.
Do not use when
- Do not use this skill for graphic design, video production, software development, or legal advice beyond the repository's stated scope.
- Do not use it when another skill in this repository is clearly more specific to the requested deliverable.
Workflow
- Collect the required inputs or source material before drafting, unless this skill explicitly generates the intake itself.
- Follow the section order and decision rules in this ; do not skip mandatory steps or required fields.
- Review the draft against the quality criteria, then deliver the final output in markdown unless the skill specifies another format.
Anti-Patterns
- Do not invent client facts, performance data, budgets, or approvals that were not provided or clearly inferred from evidence.
- Do not skip required inputs, mandatory sections, or quality checks just to make the output shorter.
- Do not drift into out-of-scope work such as code implementation, design production, or unsupported legal conclusions.
Outputs
- An AI-focused strategy, audit, system design, or prompt asset in markdown with human review and control points.
References
- Use the inline instructions in this skill now. If a directory is added later, treat its files as the deeper source material and keep this execution-focused.
<!-- dual-compat:end -->
Purpose
Design and document autonomous AI marketing agent systems that perceive their
environment, reason about what action to take, act without human prompting, and
learn from outcomes. Output is an architecture specification, a chosen workflow
template with full HITL safeguards, and a wave-appropriate implementation plan.
This skill assumes the client has completed
and has
an AI maturity wave score (1, 2, or 3). Do not recommend Wave 3 architecture
to a Wave 1 client without a phased roadmap.
Required Inputs
Ask for all of the following before generating any output:
- Client business name — trading name and legal entity if different
- Industry — sector, product or service type
- Country / city — defaults to Uganda if not specified
- Current AI maturity wave — Wave 1, 2, or 3 (from ; estimate if not available)
- Target workflow to automate — select one primary: content / sentiment monitoring / reporting / customer service / campaign optimisation
- Available technical resources — none / basic (can use no-code tools) / developer (can call APIs and self-host)
Agentic vs Generative AI: The Critical Distinction
Source: Nayebi (2025)
Most clients conflate generative AI with agentic AI. Clarify the distinction
before designing any architecture.
Generative AI is reactive.
It waits for a human prompt, generates output, then stops. Every action
requires a human to initiate the cycle. This is Wave 1 and Wave 2 behaviour.
Agentic AI is proactive.
It monitors its environment continuously, reasons about what action to take,
executes that action, and updates its behaviour based on outcomes — without
waiting for a human prompt. This is Wave 3 behaviour.
This distinction determines the architecture, the tools, the data requirements,
and the risk controls needed. A business without clean engagement data or
developer resource is not ready for a full agentic stack.
Wave guidance:
- Wave 1 clients → rule-based automation (Zapier / Make.com triggers)
- Wave 2 clients → performance-triggered AI actions using analytics data
- Wave 3 clients → full PRAL agents with continuous monitoring and learning
Most East African businesses should reach Wave 2 (Predictive ML, with 3+ months
of clean data) before building Wave 3 agents.
The PRAL Loop
Source: Nayebi (2025)
Every agentic system is built on the PRAL loop:
Perceive → Reason → Act → Learn
Map the client's chosen workflow to each stage before recommending tools.
| Stage | What the agent does | Marketing example |
|---|
| Perceive | Gathers data from its environment | Scans social mentions, reads engagement metrics, receives inbound WhatsApp messages |
| Reason | Processes data and decides what to do — using an LLM or rule-based logic | Classifies sentiment, identifies a content gap, detects a campaign underperforming |
| Act | Executes the decision | Drafts content, sends an alert, triggers a campaign boost, routes a message to a human |
| Learn | Updates its behaviour based on outcomes | Feeds performance data back into the next Perceive cycle; adjusts thresholds and templates |
Apply the PRAL loop explicitly when designing each workflow template. Label
each step so the client can see where human oversight sits.
The BDI Model for Marketing Agents
Source: Nayebi (2025)
The BDI model — Beliefs, Desires, Intentions — maps naturally to marketing
strategy and is the clearest way to define an agent's decision boundary.
| Component | Definition | Marketing application |
|---|
| Beliefs | What the agent knows | Audience data, engagement history, brand guidelines, competitor positions, product catalogue |
| Desires | What the agent is trying to achieve | Business goals — leads, awareness, retention, revenue — expressed as KPIs |
| Intentions | How the agent plans to act | Campaign tactics, content formats, channel choices, timing rules, escalation thresholds |
Prompt to use with client:
"Specify your agent's Beliefs (what data it has access to), Desires (what KPI
it optimises for), and Intentions (what actions it can take). This defines the
agent's decision boundary."
Document the BDI model before selecting any tool. An agent without a defined
decision boundary will act unpredictably.
The OODA Cycle for Real-Time Decisions
Borrowed from military strategy (Boyd, 1976), OODA is the fastest decision loop
applicable to marketing agents operating in real-time social media environments.
Observe → Orient → Decide → Act
Faster OODA cycles = competitive advantage in fast-moving social media
environments where a delayed crisis response or missed trend costs engagement.
Social listening agent example:
- Observe — scan all mentions of the brand across Facebook, Instagram, X/Twitter, and Google every hour
- Orient — classify each mention by sentiment (positive / neutral / negative / crisis) and topic category
- Decide — apply rules: respond autonomously to positive enquiries; escalate negative mentions; flag crisis keywords immediately
- Act — post pre-approved response template, or send alert to human via WhatsApp/email with full context
OODA complements PRAL: PRAL describes the agent's architecture; OODA describes
the speed and logic of its decision-making in a single cycle.
Five Agentic Workflow Templates
Select the template that matches the client's target workflow. Fully specify
the chosen template before recommending tools.
1. Content Pipeline Agent
What it does: Automates the content creation and publishing pipeline from
trend detection to post-performance feedback.
| Element | Detail |
|---|
| Trigger | Scheduled (daily/weekly) or event-driven (trending topic detected) |
| Actions | 1. Monitor trending topics and competitor content · 2. Generate draft content (caption, hashtags, image brief) · 3. Route draft to human for approval · 4. Publish approved content at optimal time · 5. Monitor post performance for 48 hours |
| HITL point | Human approves every draft before publishing — no autonomous publishing without review |
| Learn step | Performance data (reach, engagement rate, saves) fed back to refine future prompts and posting times |
| Tools | Claude API (drafting) + n8n or Make.com (orchestration) + Buffer/Hootsuite (scheduling) |
| EA feasibility | High — Wave 2 clients can implement with no-code tools |
2. Sentiment Monitoring Agent
What it does: Continuously scans social mentions, classifies sentiment, and
alerts the team when a threshold is crossed.
| Element | Detail |
|---|
| Trigger | Continuous (hourly scan) or keyword-event (brand name mentioned) |
| Actions | 1. Scan Facebook, Instagram, X/Twitter, Google reviews for brand mentions · 2. Classify mention: positive / neutral / negative / crisis · 3. Log all mentions in dashboard · 4. Alert team when negative threshold crossed (e.g., 3+ negative mentions in one hour) · 5. Suggest pre-approved response options |
| HITL point | Human selects and sends response — agent does not post responses autonomously |
| Learn step | Mis-classifications flagged by human; agent updates sentiment rules |
| Tools | Mention.com or Google Alerts (listening) + Claude API (classification) + n8n (routing) + WhatsApp Business API (alert delivery) |
| EA feasibility | High — Google Alerts + Claude API is accessible and low-cost |
3. Proactive Campaign Agent
What it does: Monitors engagement metrics and triggers a targeted response
campaign when performance drops below threshold.
| Element | Detail |
|---|
| Trigger | Metric threshold (engagement rate drops below X%, or follower growth stalls for N days) |
| Actions | 1. Pull platform analytics daily · 2. Compare against baseline benchmarks · 3. Detect underperformance · 4. Generate campaign response options (content boost, new format, re-engagement post) · 5. Present options to human for approval · 6. Execute approved option · 7. Report results after 7 days |
| HITL point | Human approves the campaign response before any content is published |
| Learn step | Successful response tactics stored; agent prioritises them in future recommendations |
| Tools | Platform analytics API + Claude API (analysis and drafting) + Make.com (orchestration) + Buffer (publishing) |
| EA feasibility | Medium — requires Wave 2 data maturity and API access to platform analytics |
4. Multi-Agent Reporting System
What it does: A team of specialised agents collaborates to produce the
monthly performance report with minimal human effort.
| Element | Detail |
|---|
| Trigger | Scheduled (last day of the month) |
| Actions | 1. Data agent — pulls platform statistics from all active channels · 2. Analysis agent — identifies patterns, anomalies, and top-performing content · 3. Writing agent — drafts narrative report with insights and recommendations · 4. Human consultant — reviews, edits, and presents to client |
| HITL point | Human reviews the full draft before delivery; no automated client-facing report |
| Learn step | Human edits tracked; writing agent refines its narrative style and recommendation quality |
| Tools | Platform APIs (data) + Claude API (analysis and writing) + n8n (orchestration) + Google Docs / Notion (output) |
| EA feasibility | Medium — high value but requires API access and developer setup for data pulls |
5. WhatsApp Response Agent
What it does: Classifies inbound WhatsApp messages, routes them to the
correct response path, and handles routine enquiries autonomously.
| Element | Detail |
|---|
| Trigger | Inbound WhatsApp Business message received |
| Actions | 1. Receive and classify message (enquiry / complaint / order / other) · 2. Route to: decision tree (simple FAQ) / Claude API (nuanced enquiry) / human agent (complaint or high value) · 3. Respond or escalate · 4. Log interaction with timestamp and classification |
| HITL point | All complaints and high-value sales enquiries routed to human immediately; agent does not resolve complaints autonomously |
| Learn step | Mis-routed messages flagged; classification rules updated monthly |
| Tools | WhatsApp Business API + Claude API (classification and response drafting) + n8n (routing logic) |
| EA feasibility | High — WhatsApp penetration in EA makes this the highest-ROI agentic workflow for most clients |
HITL Safeguard Design
Source: Nayebi (2025)
Every agentic workflow must define four safeguard components before going live.
Include this section in every workflow specification delivered to the client.
1. Autonomous decision boundary
Define what the agent can decide and act on without human review. Limit this to
decisions that are: low-risk, routine, reversible, and within a defined value
threshold (e.g., scheduling a post, classifying a mention, logging a message).
2. Escalation triggers
Define what forces the agent to stop and wait for a human. Escalation is
mandatory when a decision is: high-risk, irreversible (e.g., publishing to
public), sensitive (crisis keywords, complaints, legal mentions), or above a
value threshold (e.g., enquiry worth over UGX 500,000).
3. Escalation mechanism
Specify: how the human is alerted (WhatsApp message, email, Slack), what
information they receive (full context, agent's recommended options), the
expected response time, and the override protocol if no response is received.
4. Audit trail
Every agent action must be logged with: timestamp, action taken, data that
triggered the action, reasoning or rule applied, and outcome. This log is
reviewed monthly to improve agent performance and demonstrate accountability.
Three-Wave Implementation Roadmap
Match the roadmap recommendation to the client's current wave.
| Wave | Readiness criteria | What to build | Effort |
|---|
| Wave 1 — Automation | No analytics data required; any technical level | Zapier or Make.com automations that trigger AI content drafts on a schedule. Rule-based, no learning, no API calls. | 1–2 days setup |
| Wave 2 — Performance-triggered | 3+ months of clean engagement data; basic technical resource | Connect analytics data to AI for performance-triggered actions (e.g., engagement drop → draft new content). Requires platform data export or basic API access. | 1–2 weeks setup |
| Wave 3 — Full agentic | Clean data, developer resource, HITL safeguards in place | Full PRAL agents with continuous monitoring, LLM reasoning, and feedback loops. Requires API access, self-hosted orchestration (n8n), and ongoing maintenance. | 4–8 weeks minimum |
Do not propose Wave 3 to a Wave 1 client without a phased roadmap that moves
them through Wave 2 first.
Tool Stack Options
Recommend tools based on the client's technical resources and budget.
| Tool | Role in agentic stack | EA accessibility | Approx. cost |
|---|
| Claude API | LLM reasoning layer — classification, drafting, analysis | Yes — API account required | Pay-per-token |
| n8n | Workflow orchestration (self-hostable, open source) | Yes — developer resource needed for self-hosting | Free (self-hosted); from $20/month (cloud) |
| Zapier AI | No-code workflow automation with AI steps | Yes — browser only, no dev required | Free tier; from $19.99/month |
| Make.com | Visual no-code workflow builder | Yes — browser only, no dev required | Free tier; from $9/month |
| Hootsuite / Buffer | Publishing and scheduling layer | Yes — widely used in EA | From $15/month |
| Brandwatch / Mention | Social listening layer for sentiment monitoring | Limited — pricing is a barrier for small clients | From $99/month |
| WhatsApp Business API | Inbound message routing and response | Yes — high EA penetration; via Meta or third-party | From $0 (first 1,000 conversations/month free) |
For Wave 1 clients with no technical resource: Zapier or Make.com + Claude
(via ChatGPT or Claude.ai interface, not API) is the most accessible entry point.
For Wave 3 clients with developer resource: n8n (self-hosted) + Claude API is
the recommended EA-feasible stack.
Quality Criteria
Output meets standard when it satisfies all of the following:
- Client's current AI maturity wave is identified and a wave-appropriate architecture is recommended — no Wave 3 proposal for a Wave 1 client without a phased roadmap
- At least one agentic workflow template is selected and fully specified: trigger, PRAL mapping, actions, HITL point, learn step, and tools
- The PRAL loop is explicitly mapped for the chosen workflow — each stage labelled
- The BDI model is documented: Beliefs (data sources), Desires (KPI optimised), and Intentions (actions the agent can take)
- HITL safeguards are defined: autonomous decision boundary, escalation triggers, escalation mechanism, and audit trail
- Tool stack is recommended based on the client's technical resources and budget with EA accessibility noted
- EA feasibility is assessed — WhatsApp-based and no-code workflows prioritised for clients without developer resource
References
- Nayebi, F. (2025) Foundations of Agentic AI for Retail. Gradient Divergence.
- Venkatesan, R. and Lecinski, J. (2026) The AI Marketing Canvas, 2nd edn. Stanford University Press.
- Farri, E. and Rosani, G. (2025) HBR Guide to Generative AI for Managers. Harvard Business Review Press.