I have a working-but-fragile Chrome extension that crawls pages and “dorks” (niche search queries) looking for HTTP API tokens that match a specific regex. The job is twofold: 1. Repair and streamline the current codebase so the scanner reliably captures and locally stores every hit. Right now it sporadically misses matches and sometimes stops writing to storage. 2. Enrich the detection logic with fresh, private dorks and widen the sweep to cover the full set of indicators I care about: Keywords, Phrases, API endpoints, HTTP API strings and any related variations you can dream up. The flow I picture: • Content script injects, scans DOM + network responses in real time. • Regex engine identifies tokens, deduplicates them and saves to local storage / optional webhook. • Popup or background page shows a running list with export to CSV. Feel free to suggest smarter techniques—service-worker listeners, indexedDB for larger caches, or a lightweight backend if you think it helps. Just keep everything inside the Chrome extension ecosystem so installation stays one-click. I’ll supply the existing source, token regex, and a starter list of public dorks. Your deliverable is the fixed, upgraded extension plus a curated set of fresh private dorks ready to drop in. If you’ve built security, scraping, or OSINT tools before, tell me how you’ll plug that experience into this project and the specific libraries or Chrome APIs you plan to rely on (e.g. chrome.scripting, declarativeNetRequest, fetch interception). Looking forward to seeing how you can push this further.