Instructions
A guide to configuring and running SEO crawls with Screaming Frog Cloud. All crawl settings are managed through .seospiderconfig files created in Screaming Frog Desktop.
1. Create a Configuration File
All crawl settings (threads, URL limits, extraction rules, user agent, etc.) are controlled by a .seospiderconfig file. Create one in Screaming Frog Desktop:
- Open Screaming Frog SEO Spider on your desktop
- Configure your crawl settings (see sections below for key settings)
- Go to File > Configuration > Save As
- Save the
.seospiderconfigfile
2. Key Settings to Configure in SF Desktop
URL Crawl Limit
Configuration > Spider > Limits > Crawl Limit
Set the maximum number of URLs to crawl. Without this, SF will crawl the entire site which can take hours for large sites. Recommended: 500 for quick audits, 5000 for full audits.
| Setting | Where in SF Desktop | Recommendation |
|---|---|---|
| Crawl Limit | Configuration > Spider > Limits | 500 (quick) / 5000 (full) |
| Threads | Configuration > Speed | 10-15 recommended |
| JS Rendering | Configuration > Spider > Rendering | OFF unless crawling SPA (React/Angular/Vue) |
| URL Filtering | Configuration > Include / Exclude | Use to focus on specific sections (e.g. /blog/) |
| User Agent | Configuration > User-Agent | Googlebot or Chrome |
| Respect Robots.txt | Configuration > Robots.txt | ON for production, OFF for full audit |
3. Custom Extraction Rules
Extract specific content from every crawled page. Configure in SF Desktop:
- Go to Configuration > Custom > Extraction
- Click Add to create a new rule
- Choose the selector type (CSS Selector, XPath, or Regex)
- Enter the selector pattern
- Save the config file
Tip: Inspect the target site first (right-click → Inspect Element) to find the right CSS selectors for the content you want to extract.
Common extraction examples:
| Goal | Type | Pattern | Extract |
|---|---|---|---|
| Page body content | CSS Selector | .entry-content | Extract Text |
| Main content area | CSS Selector | main, article, #content | Extract Text |
| Product price | CSS Selector | .product-price, .price | Extract Text |
| Schema markup | CSS Selector | script[type="application/ld+json"] | Inner HTML |
| Email addresses | Regex | [a-zA-Z0-9._%+-]+@[a-zA-Z0-9.-]+\.[a-zA-Z]{2,} | Extract Text |
4. Upload Config and Start Crawl
- Go to New Crawl from the sidebar
- Enter the target URL
- Upload your
.seospiderconfigfile (optional — default config is used if not uploaded) - Review the parsed settings and extraction rules displayed below the upload
- Click Save Config to save it for reuse
- Click Start Crawl
Previously saved configs can be loaded via the "Load Saved Config" link.
5. Viewing Results
- Pages tab — Summary stats: total pages, status codes, errors, issues
- Issues tab — SEO issues detected with severity, count, and downloadable CSV per issue
- SF Exports tab — Download raw Screaming Frog CSV files
- Custom extraction data is automatically cleaned and stored in the database. Raw CSV is also available in SF Exports
6. Managing Config Files
- Save — Upload and save configs for reuse across crawls
- Load — Select a previously saved config when starting a new crawl
- Download — Download a saved config to edit in SF Desktop and re-upload
- Configs store all SF settings including extraction rules, URL limits, speed, rendering, and filtering
7. Tips for Faster Crawls
- Always set a URL crawl limit in SF Desktop to avoid crawling entire large sites
- Keep JS Rendering OFF unless crawling a SPA
- Use URL filters to focus on specific sections (e.g.
/blog/) - Increase threads to 15-20 for faster crawls on robust servers
- Disable image/CSS/JS checking in SF Desktop if you only need HTML page data