Guide intermediate

n8n's AI Agent nodes in production: a 30-day operator review

We ran two real customer-support workflows on n8n's AI Agent nodes for 30 days. Here are the three things that broke, the patches we shipped, and the cost data.

Published May 5, 2026 by Pondero Editorial

This article contains affiliate links — disclosure.

Table of Contents
Pondero, operated by Hildebrandt AI LLC, earns a commission from some links on this page. This does not influence our editorial decisions. Read our affiliate disclosure

n8n’s AI Agent nodes in production: a 30-day operator review

Published May 5, 2026 by Pondero Editorial

Hand-drawn illustration of n8n AI Agent nodes in production
n8n's AI Agent node wraps LangChain. The wrapper is the easy part; the production failures live underneath.

In short

n8n ships native AI Agent nodes that wrap LangChain agent loops inside its visual workflow editor. We ran two real customer-support workflows on n8n agents for 30 days, processing roughly 12,000 workflow runs across both. The verdict: production-ready for ops teams who already run n8n and can debug a LangChain agent loop. Not yet production-ready for teams who want a no-code agent platform with managed guardrails. Three things broke (tool-call loops, memory blowups, OpenAI rate limits) and all three had patches we shipped within the 30 days. If you are picking between n8n agents, Lindy, and Zapier Agents, this review tells you when each one fits.

What n8n’s AI Agent node actually does

n8n’s AI Agent node is a wrapper around LangChain’s agent abstractions, exposed as a draggable node in the n8n visual editor. You connect it to a chat model node (OpenAI, Anthropic, Ollama, etc.), a memory node (buffer, window, summary), and any number of Tool nodes (HTTP request, code execution, custom workflows, MCP).

The agent loop runs inside the n8n execution engine. Each iteration: the model sees the conversation, picks a tool to call (or replies directly), the tool executes, the result feeds back to the model, repeat until the agent decides it is done or hits an iteration cap.

n8n is open-source under the Sustainable Use License (a fair-code license, not strictly OSI-approved). The agent code lives in the LangChain integration package and follows LangChain’s release cadence. For the latest n8n releases, see github.com/n8n-io/n8n/releases.

Agent vs Tool nodes vs Chain nodes

n8n’s AI category has three distinct node families that confuse first-time users:

  • Chain nodes: linear LLM calls with no tool use. Input goes in, prompt template applies, model responds, output flows down. Use for summarization, classification, single-shot generation.
  • Tool nodes: wrappers that expose any n8n integration (HTTP, Postgres, Slack) to an Agent or Chain. Tools do not run on their own; they wait to be called.
  • Agent nodes: the iterative loop. Picks tools, calls them, reasons over results.

If your task is “translate this text” or “categorize this email,” use a Chain. If your task is “look at this email, decide whether to refund or escalate, and write a reply,” use an Agent.

Two production workflows we built

Customer support triage workflow on n8n agent nodes
The triage workflow: inbound email enters, agent picks tools, drafts reply, queues for human approval.

Customer support triage with HubSpot tools

Trigger: inbound email from our support inbox.

Agent prompt (abbreviated): “You triage support tickets. Categorize the ticket. Look up the customer in HubSpot. If the customer is on the Pro plan, draft a high-touch reply. Otherwise, draft a templated reply. Always queue for human approval.”

Tools available to the agent: HubSpot contact lookup, HubSpot deal lookup, knowledge-base search (Algolia), draft-reply HTTP call, approval-queue HTTP call.

Volume: ~7,000 tickets across 30 days. Roughly 80% triaged correctly on first pass; 12% required a category override at the human-approval step; 8% were escalated to engineering or refund teams.

Internal RAG over Confluence docs

Trigger: Slack /ask-docs command.

Agent prompt: “Answer engineering questions using only our Confluence documentation. Cite the page you found the answer in. If the docs do not contain the answer, say so.”

Tools: Confluence search, Confluence page fetch, Slack reply.

Volume: ~5,000 queries across 30 days. Roughly 70% were answered with a source citation; 18% returned “the docs do not contain that”; 12% returned a hallucinated answer that an engineer flagged.

What broke (and what we did about it)

Tool-call loops on ambiguous inputs

Symptom: agent would call HubSpot lookup, get an empty result, call it again with a slightly different query, get empty again, repeat for 10 iterations until hitting the cap.

Root cause: the agent treated empty results as “try harder” instead of “this customer is not in HubSpot.”

Patch: we added an explicit “if a lookup returns empty, treat the customer as new and proceed without HubSpot context” instruction to the system prompt. Loop frequency dropped from roughly 8% of runs to less than 1%. Still not zero, but tolerable.

Memory window blowups with long threads

Symptom: long support threads (10+ back-and-forth emails) would push the conversation past the model’s context window. The agent would either truncate aggressively (losing key context) or fail with a token-limit error.

Root cause: we started with a buffer-memory node that kept the entire thread. That works fine for short threads and breaks at scale.

Patch: switched to a summary-memory node that keeps a running summary of older turns and the verbatim text of recent turns. Token usage per agent invocation dropped roughly 40% on long threads, and the agent kept the relevant context.

OpenAI rate-limit handling

Symptom: during morning email surges, OpenAI’s tier-2 rate limits (3,500 RPM) were hit. n8n’s default behavior was to fail the workflow run.

Root cause: no retry logic on rate-limit errors out of the box.

Patch: wrapped the chat model node in an n8n error-trigger workflow that catches 429 responses and retries with exponential backoff. We also moved high-volume tasks (the RAG lookups) to Claude Haiku via the Anthropic node, which split the rate-limit pressure across two providers.

For OpenAI and Anthropic API pricing, verify at openai.com/api/pricing and anthropic.com/pricing. Both have shipped pricing changes in 2026.

Cost data: 30 days, ~12k workflow runs

Across both workflows, our model spend looked like this:

Cost lineAmount
OpenAI API (GPT-4o for triage)~$840
Anthropic API (Claude Haiku for RAG)~$190
n8n Cloud (Pro tier)$50
Total monthly run-rate~$1,080

Per-ticket cost on the triage workflow: about $0.12. Per-query cost on the RAG workflow: about $0.04. Both numbers include only model spend; n8n hosting is a fixed line.

For comparison, our Lindy pilot in March cost roughly $0.20 per triage ticket at the same volume but required less in-house engineering time to set up.

When to use n8n agents vs Lindy vs Zapier Agents

n8n agents vs Lindy vs Zapier Agents
n8n is the engineer's pick. Lindy is the no-code pick. Zapier Agents fits everyone else.

Use n8n agents if:

  • You already run n8n and have a developer who can debug a LangChain loop
  • You need self-hosting (n8n Community Edition is free; n8n Cloud is the managed option)
  • You want fine-grained control over memory, tools, and agent prompts

Use Lindy if:

  • You want a no-code agent platform with built-in approval gates
  • Your team is sales ops or revops, not engineering
  • You are willing to pay roughly 60% more per task for less setup time

Use Zapier Agents if:

  • You already pay for Zapier and want agent capabilities without a new vendor
  • Your workflows involve apps that Zapier integrates with deeply (HubSpot, Salesforce, Gmail)
  • Your task volume is moderate; per-task pricing rewards lower volume

For a deeper Lindy walkthrough, see Lindy for sales ops: a 30-day rollout. For the wider field, see best AI automation tools for ops leads.

What we cannot tell you

This review is two workflows, 30 days, one team. We have not run a five-team comparison or a year-long longitudinal study. Your tool-call loop frequency, memory blowup rate, and rate-limit pressure will depend on your traffic shape, your model choice, and your tools’ latency characteristics. Treat the cost numbers as a baseline; expect variance.

FAQ

Is n8n really free? n8n Community Edition is free to self-host. n8n Cloud is the managed paid tier. Verify current Cloud pricing at n8n.io/pricing.

Can n8n agents call MCP servers? Yes. n8n added MCP node support in early 2026. We covered the workflow tools angle in n8n MCP workflow tools.

What’s the iteration cap? Configurable per agent node. Default is 10. We run 15 on the triage agent and 8 on the RAG agent.

Does n8n support local models? Yes via the Ollama node. Quality drops vs. hosted models; latency is better in some setups.

Verdict

n8n’s AI Agent nodes are production-ready for engineering teams who already run n8n. They are not yet production-ready for no-code ops teams who want managed guardrails out of the box. Our two workflows survived 30 days at ~12k runs with three patches and a $1,080 monthly run-rate. We are keeping both in production and recommend n8n agents as the engineer’s pick in the agent platform space.

Try n8n. Start with a Chain node, graduate to an Agent node when you need iterative tool use.


Related: n8n tool page · n8n MCP workflow tools · Lindy for sales ops · Best AI automation tools for ops leads