I’m having an issue with the “HTTP Request” node for web scraping. I’m running n8n on a server with the latest self-hosted version. After around 15 executions, the node starts running indefinitely. The timeout doesn’t seem to be applied, and as a result, it blocks the proper execution of the rest of my workflow. I’ve tried several solutions, but none of them worked.
Not individually, everything works fine. The problem occurs after about 15 loops, and then it runs indefinitely. and it happens with many different URLs; this is just the last example where it crashed
Can you provide the list of pages to run sequentially to trigger the issue?
Or is it the list you mentioned earlier? If so, the list has bad urls and images, is that expected?
I just tested my workflow with only the bad URLs and only the image URLs, and it worked perfectly fine.
Could the problem be related to cache or memory?
Depending on your circumstance and the solution you need, you might need to rethink this. It seems the problem you’re faced with is literally overloaded memory. Just like RAM in a computer system, or short term memory in our human brains, we need to intelligently re-sort, organize and prioritize all the bits for longer-term usage. (Memory Manager Node)
You might consider adding a knowledge base/RAG tool and iteratively store the data you need to the knowledge base to reduce the burden on the memory.