I’ve drafted a detailed technical brief for an MVP that will collect construction-sector leads in Italy and Spain without tripping Google Maps’ 120-result ceiling. The flow I have in mind is: • Automatically fan out searches by city or ZIP code (“geo-splitting”) so every company appears in the results. • Capture the company name, address, phone number and any email Google exposes. • Follow the website link, harvest additional emails plus the VAT number that rarely shows on Maps. • Pass the page text through the OpenAI API and return a concise, SEO-friendly description we can drop straight into our CRM. • Export everything to CSV/JSON and, ideally, feed the file to n8n for downstream automations. I’m comfortable with Apify as the scraping backbone and either n8n or straight Python for the orchestration logic, but I’m open to suggestions if you can hit the same milestones faster or more robustly. Intermediate-level scraping expertise is plenty; what matters most is clean, maintainable code and a repeatable deployment process—no brittle, one-off scripts. If this sounds like a fit, I’ll share the requirement document so you can gauge scope and quote accurately. Looking forward to your thoughts and, if all checks out, a quick start.