Hey everyone!
So, I have this big school project where I need to grab product details from a bunch of different websites. Basically, I’m trying to look up specs using either the product name or the EAN code (that barcode number).
The tricky part is, I have a huge list—like, 300 to 1,000 products—and I need to do them all at once, not one by one (I’d be here forever ).
I need to get a really clean list of these specific specs for each product:
EAN number
Depth, Height, and Width
Color and Material
Seat depth, Seat height, and Seat width
Armrest height
Backrest height
Maximum load capacity
How many people it holds
Does it swivel? (Yes/No)
Product weight
Has anyone here done anything like this before, maybe using some kind of scraping tool or script? Any tips on the best way to handle this big batch of lookups would be a lifesaver!
Describe the problem/error/question
What is the error message (if any)?
Please share your workflow
(Select the nodes on your canvas and use the keyboard shortcuts CMD+C/CTRL+C and CMD+V/CTRL+V to copy and paste the workflow.)
Hey @logidy The best way to handle your massive product spec retrieval task getting data for 300 to 1,000 products all at once is by setting up an automated system. This is much faster and more reliable than trying to copy-paste.
You have two main technical options, both using Python:
1. For simple, static websites: Use a library like Scrapy to quickly grab data. It’s designed specifically for high-speed, large-scale scraping, letting you process many products in parallel.
2. For complex, dynamic websites (like modern e-commerce stores): Use a headless browser tool like Playwright. This simulates a real web browser, allowing it to “see” and interact with all the JavaScript elements that load the specifications, even if they aren’t visible right away.
To manage the list of 1,000 products and the final output, you’d integrate your scraping script with a workflow automation tool like n8n. You’d start by putting all your product EANs/titles into a Google Sheet. The n8n workflow would read this list, feed each product into your scraping script (Scrapy or Playwright), and then automatically write the clean, extracted data (Depth, Color, Max Load, etc.) back into the corresponding rows of your Google Sheet.
This method handles the batch processing and actuallly ensures your final data is neatly organized and standardized.