Hello
I’d like totest if an image is rotated or not.
I’ve tried multiple solutions, but still stuck as I can identify if an image is not correctly oriented, but I can’t apply some code (JS or Python) to rotate it, then save it (I’m not sure where yet, but I could use a SFTP node, a temporary execute command node, then to send it in a directory)
I can’t make the binaries available and / or also loosing the URL passed to the llm.
I have tested the OpenAI llm , with URLs binaries, or binaries, and also Anthropic node (that needs biaries only).
Last issue, with Anthropic I may only pass JPG files, as the Mime Type is not recognized correctly except for JPG. ( and they claim to accept webp, png etc… also)
What is the error message (if any)?
Please share your workflow
I’m going to add the workflows after posting as I have 403 errors.
It’s an interesting problem because without an original to compare, who is to say if an image has really been “rotated”? And seriously, what does “up” and “down” even mean for an AI?
I created a quick test below which is able to identify if the image is “rotated” but is unable to correctly provide the params to rotate it. I would imagine specialised AI vision models to do a lot better than Multimodal but I haven’t experimented enough to say for certain.
Yes it does work !
I think I’ve mixed up a bit the versions I was working on, but in the prompt if you stipulate that it has to be easily human readeable horizontally, it takes it !
I have successfully tested this approach.
What’s important is to take things step by step, with multiple llm calls, with the lowest tokens possible, produce a json, sorts file that needs to be rotated, then rotate them (here I fail atm to use a tool for that) then keep going on analysis…etc…
Bascially, reducing the “noise” at each step.
I’ll share some stuff very soon, I guess we’ll find out very soon