Hi all,
I’m currently exploring how to work with image inputs in gpt-image-1 using n8n, and had a use case I wanted to validate:
Goal:
I’d like to provide multiple character images (e.g., 3–5 separate portraits of characters) and have the AI generate a single composite image — something like a game cover or poster that includes all of them, properly arranged/stylized.
What I Know:
- This can be done using the OpenAI API in Python by uploading multiple image references and prompting GPT to create a merged visual output.
- I can work with OpenAI nodes or the HTTP Request node in n8n to hit their API endpoints.
My Questions:
- Is it currently possible to do multi-image input → single image output using OpenAI’s image generation tools directly within n8n?
- If yes, what’s the best workflow for handling multiple image inputs? Would this involve storing them in a binary format, referencing image URLs, or something else?
- Has anyone done something similar using gpt-image-1, or similar for multi-character composition workflows in n8n?
Any guidance, templates, or best practices would be hugely appreciated!
Thanks in advance