Handling Long-Running Workflows with Native MCP + Controlling AI Response Behavior

Hi everyone!

I’m building a chatbot that uses n8n’s native MCP to expose workflows as tools. I’ve run into two challenges and would love to hear how (or if) others have tackled them.

Setup

  • n8n with native MCP Server Trigger

  • Workflows exposed as MCP tools (e.g., “Generate LinkedIn Post Ideas”)

  • Client: ChatGPT (via MCP)

Challenge 1: Long-Running Workflows (60+ seconds)

Part 2 of my workflow could take 2-3 minutes to complete (they involve multiple API calls, AI processing, etc.). I’m hitting the 60-second MCP timeout limit.

My current workaround: I’m using an async pattern where:

  1. The workflow immediately returns a “processing started” message

  2. Results are delivered via polling (Another MCP Command that will retrieve the post from the DB)

Question: Is there a cleaner built-in way to handle this?

Challenge 2: AI Client Adds Its Own Content to Tool Results

This is the trickier one. When I ask ChatGPT to “Generate LinkedIn post ideas” (which triggers my n8n workflow), I get back:

  • :white_check_mark: The actual workflow output (my generated ideas)

  • :cross_mark: PLUS ChatGPT’s own generated ideas on top of it

The AI seems to interpret the tool call as “user wants LinkedIn ideas” and tries to help by generating more, rather than just relaying what the tool returned.

What I want: The AI should only return the workflow output, not add its own content.

What I’ve tried:

  • Adding an edit node at the end mentioning: do not return anything

  • Adding system prompts like “Only relay tool outputs, do not generate additional content”

Neither fully solved it.

Question: Has anyone found a reliable way to make the AI client (ChatGPT, Claude, etc.) just return the MCP tool output without adding its own interpretation/content?

Is this something that needs to be handled on the n8n side (in how the response is formatted), or is it purely a client-side prompt engineering challenge?

For Challenge 1, your async pattern with polling is the right approach. MCP has hard timeout limits, so immediate return plus polling is standard for long workflows. No built-in alternative exists yet.

For Challenge 2, the AI always wants to be helpful and wraps tool outputs. Try returning results as structured JSON with a clear instruction embedded in the response itself like “Final result - do not modify or add content”. Also prefix your system prompt with “You are a passthrough assistant” to reinforce behavior.

This is client-side behavior, not n8n. Different models handle it differently.

This topic was automatically closed 90 days after the last reply. New replies are no longer allowed.