Promote your business with Collaborator.
9136 unique sites to publish the articles.

Screaming Frog Seo Spider: The Ultimate Beginer's Guide

Table of contents

Screaming frog SEO Spider is a website SEO audit tool. We've compiled a comprehensive list of best practices for 2022 on how SEOs, PPCs, and digital marketers can use Screaming frog to streamline their workflow.

When we wrote the instructions, we analyzed what most users are looking for in this program:

  • Screaming Frog how to download for free 
  • What is Screaming Frog SEO Spider and what is it used for?
  • How to run a full site scraping in Screaming Frog SEO?
  • How to use Screaming Frog to improve on-page SEO?
  • How to do a technical audit of the site using Screaming Frog SEO?
  • How to use the Screaming Frog SEO Spider tool to check internal and backlinks? 
  • What about the Screaming Frog pricing and is it free to use?

These and many other questions will be answered in the article. This is a real the Ultimate Beginer's Guide Prepared with the latest updates

New 16th version, codenamed «Marshmallow», Screaming Frog SEO Spider released in September 2021 

1

How to run a full site scraping in Screaming Frog SEO

How to crawl the full site

Before launching the full site scraping in Screaming Frog SEO, determine:

  • what information do you want to get, 
  • how big the site is,
  • what part of the site do you need to check for completeness. 

In the case of large sites, it is better to limit the analyzed area to a subsection of URLs in order to get a representative sample of the data. This keeps file sizes and export data more manageable. We will return to these details later.

To fully scrape the site with all subdomains, you will need to change the settings a bit before starting. This is because by default Screaming Frog checks only the subdomains you have entered. Other additional ones will be treated as external links by the algorithm. To check additional subdomains, you need to change the settings in the Configuration menu. By selecting Spider > Crawl all Subdomains, you will ensure that the algorithm crawls any links to subdomains of your site.

Step 1 - Configuring Domain and Subdomain Scraping

screeming frog

Step 2 - configuring the whole site crawl

If you started crawling from a specific subfolder and still want Screaming Frog to crawl the whole site check «Crawl Outside of Start Folder».

sreaming frog

Pro Tip — Exclude image, CSS, and Javascript crawling 

To disable crawling of images, CSS, and Javascript files in Screaming Frog uncheck the items you don't need. This will save time and disk space (reduce the size of scanned files).

How to set up Screaming Frog to crawl a single folder 

To restrict the Screaming Frog Seo Spider to a single folder just enter the URL and hit start without changing any of the default settings. If you have overwritten the initial settings, reset the default configurations in File > Configuration > Clear Default Configuration. See instructions below. 

screming frog

If you want to start crawling in a specific folder, but continue in the remaining subdomain, select «Crawl outside of start folder» from the Configuration menu before entering the start URL. This screenshot will help you understand the settings. 

How to check subdomains and subdirectories: customizing Screaming Frog

To restrict the Screaming Frog check to a specific set of subdomains or subdirectories, you can use RegEx to set these rules in the Include or Exclude settings in the Configuration menu.

Exception

In this example, we have crawled every page on seerinteractive.com, excluding the "About" page in each subdomain.

Step 1 — write the exclusion rules

Open Configuration > Exclude; use a wildcard expression to identify the URLs or parameters you want to exclude.

screaming frog

Step 2 — Testing how the exclusion will work

Test the expression to make sure there are pages that should be excluded before starting the test.

screaming frog download

Inclusion

In the example below, we wanted to crawl a command subfolder on seerinteractive.com. Use the Test tab to test multiple URLs to make sure the RegEx for the sub-rule is set correctly.

This is a great way to test large sites. In addition, Screaming Frog recommended this method if you need to split and validate a large domain.

site crawl

How to collect all pages on a site

By default Screaming Frog checks all images, JavaScript, CSS, and flash files.

To collect all the pages of the site (HTML only) in Screaming Frog uncheck «Check Images», «Check CSS», «Check JavaScript» and «Check SWF» in the Configuration menu.

screening frog

When you disable the above options, Screaming Frog Seo Spider will collect all the pages on the site with internal links to them.

Once the crawl is complete, select the «Internal» tab and filter the results by «HTML». Click «Export» and you will have a complete list of pages in CSV format. See instructions on how to use this feature. 

spider frog

Pro Tip — Save the Screaming Frog settings you use for each check

seo frog

How to list all pages in a separate subdirectory

In addition to resetting «Check CSS», «Check CSS», «Check JavaScript», and «Check SWF», you can also reset «Check links outside folder» in settings. If you disable these settings and crawl, you will additionally get a list of all pages in the start folder.

How to Find All Subdomains on a Website Using Screaming Frog and Check External Links

There are several ways to find all subdomains on a website.

Method 1: use Screaming Frog to get all subdomains on the specified site.

Go to Configuration > Spider and make sure the «Crawl all Subdomains» option is selected. This will help check any related subdomain during the crawl. Note that unlinked subdomains will not be found.

Method 2: use Google to find all indexed subdomains.

Using the Scraper Chrome extension and advanced search operators, you can find all indexed subdomains for a given domain.

Step 1 — using the site and inurl operators

Start by using the site search operator: in Google to limit your results to a specific domain. Then use the -inurl search operator to narrow your search results by removing the main domain. After that, you will start to see a list of Google-indexed subdomains that do not contain the main domain.

Step 2 — use the Scraper extension

Use the Scraper extension to extract results into a Google spreadsheet. To do this, right-click on the link in the SERP, select «Scrape similar» and export to a Google Doc.

Step 3 — remove extra data

In Google Docs use the following function to trim URLs in a subdomain:

=LEFT(A2,SEARCH("/",A2,9))

This guide will help you remove any subdirectory, page or filename at the end of your site. The function tells the sheets or Excel to return whatever is to the left of the slash. The leading number 9 is significant because we are asking to start looking for a slash after the 9th character. This is taken into account for the https:// protocol with a length of 8 characters.

Remove duplicates from the list and load it into Screaming Frog in list mode.

screaming frog alternative

You can insert a list of domains manually, use a function or upload a CSV file.

crawl site screaming frog

Method 3: find subdomains in Screaming Frog 

Enter the domain's root URL into the tools, which will help you find sites on the same IP address. Alternatively, use dedicated search engines such as FindSubdomains. Create a free account to login and export the list of subdomains. Then upload the list to Screaming Frog using list mode.

Upon completion of the crawl, you will be able to see the status codes, as well as any links to subdomain home pages, anchor texts and duplicate page titles etc.

How to analyze an online store and other large sites using Screaming Frog Spider

What you need for scraping large sites using Screaming Frog Spider: 

  • you can increase the memory allocation.
  • split the check into subdirectories or check only specific parts of the site using the Include/Exclude settings.
  • you can choose not to select images, JavaScript, CSS and flash files. By deselecting these options in the Screaming Frog Seo Spider settings, you save memory for crawling HTML only.

Pro Tip — pause crawling when necessary

Until quite recently, Screaming Frog SEO Spider could pause or crash crawling on a large site. Now, with default database storage, you can pick up the crawl where you left off. In addition, you have access to a queue of URLs in case you want to exclude or add additional parameters to test a large site.

screaming frog

How to check a site that is stored on an old server

In some cases, old servers may not be able to process a certain number of requests per second by default. By the way, we recommend setting this limit in order to respect server etiquette. It's best to let the client know when you plan to check the site against unknown user agents. In this case, they may whitelist your IP before you start crawling the site. Otherwise, you will send too many requests to the server and accidentally crash the site.

To change the crawl speed, open Configuration > Speed ​​and in the pop-up window select the maximum number of threads that should run simultaneously. In this menu, you can also select the maximum number of URL requests per second.

screamingfrog

Pro Tip — increase Response Timeout in settings

If the crawl result gives a lot of server errors, go to Configuration > Spider > Advanced in settings and increase «Response Timeout» and «5xx Response Retries» to get better results.

screaming frog seo spider

How to analyze a site that requires cookies

Screaming Frog Seo Spider allows you to crawl sites that require cookies. Although search bots don't accept cookies, if you're crawling a site and need to allow them, just go to Configuration > Spider > Advanced and select «Allow Cookies».

screaming frog seo spider

Can Screaming Frog check pages that require authentication?

Yes, you can manage authentication in the Screaming Frog settings. When the Screaming Frog algorithm encounters a password-protected page, a pop-up window will appear where you can enter the required username and password.

Forms-based authentication is a powerful feature that may require JavaScript to work effectively.

Note that forms-based authentication should be used infrequently and only by advanced users. The algorithm is programmed so that it clicks on every link on the page. So this could potentially affect links that create posts or even delete data.

To manage authentication, go to Configuration > Authentication.

To turn off authentication requests go to Configuration > Authentication and uncheck «Standards Based Authentication».

screaming frog pricing

2

Search for internal and external links in Screaming Frog

How to get information about all internal and external links on the site

If you do not need information about JavaScript, CSS and flash files, disable these options in the settings menu to save processing time and memory.

what is screaming frog

Once the algorithm has finished crawling, use «All links» in the Bulk Export menu to export the CSV. You will learn about the location of all links along with the corresponding anchor texts, directives etc.

All links can be a big report. Keep this in mind when exporting. For large sites the export may take time.

To quickly count the number of links on each page open the «Internal» tab and filter by «Outlinks». Anything above 100 may need to be revised.

how to use screaming frog

How to find broken links in Screaming Frog (page or website)

Screaming Frog Seo Spider allows you to find broken pages. Once the algorithm has finished checking, filter the results in the «Internal» tab by «Status Code». All 404, 301 and other status pages will be shown.

By clicking on any individual URL as a result of the check, you will see the information change at the bottom of the program window. By clicking on the «In Links» tab at the bottom of the window, you will find a list of pages that are linked to the selected URLs, along with the anchor text and directives used on those links. You can use this feature to monitor pages where internal links need to be updated.

To export a complete list of pages with broken or redirected links, go to Bulk Export > Response Codes and select «Redirection (3xx) In Links», «Client Error (4xx) In Links», or «Server Error (5xx) In Links». This way you will get the exported data in a CSV file.

To export a complete list of broken pages, open the Bulk Export menu. Scroll down to the response codes and look at the following reports:

  • no response inlinks;
  • redirection (3xx) inlinks;
  • Redirection (JavaScript) inlinks;
  • redirection (meta refresh) inlinks;
  • client error (4xx) inlinks;
  • server error (5xx) inlinks.

what does screaming frog do

Screaming Frog Seo Spider finds all broken links on the site. Looking at these reports should give you an idea of ​​which internal links need to be updated. This will ensure that they point to the canonical version of the URL and are distributed efficiently.

How to use Screaming Frog SEO Spider to check outbound links

Screaming Frog SEO Spider allows find external links from the site you are checking.

Make sure the «Check External Links» option is checked in Configuration > Spider.

After the algorithm has finished checking, go to the «External» tab in the top window. Then filter the data by «Status Code» and you will see URLs with all codes except 200. By clicking on a specific URL in the search result and then on the «In Links» tab at the bottom of the window, you will find a list of pages that indicate the selected URL. You can use this feature to identify pages that need to update outbound links.

To export a complete list of outgoing links, go to Bulk Export > External Links.

what does screaming frog do

For a complete list of all locations and anchor texts for outbound links, Bulk Export > All Outlinks. This report also contains outgoing links to your subdomain. If you want to exclude your domain, refer to the «External Links» report above.

How to find redirected links

After the algorithm has finished checking, select the «Response Codes» tab in the main interface and sort by Status Code. Because Screaming Frog uses regular expressions to search, validate the following filter criteria: 301|302|307. This should give you a fairly large list of all the links that came back with a certain type of redirect.

It will also let you know if the content has been permanently moved, found or redirected, or temporarily redirected according to the HSTS settings (most likely caused by the 307 redirect in Screaming Frog). Sort by «Status Code» and you can split the results by type. Click on the «In Links» tab at the bottom of the window to see all pages that have redirected links.

If you export directly from this tab, you will only see the data shown at the top of the window (original URL, code status, and where the redirect went).

To export the full list of pages with redirected links, go to Bulk Export > Response Codes and select «Redirection (3xx) In Links». You will receive a CSV file containing the location of all redirected links. To display only the internal redirect, filter the «Destination» column in the CSV to include only your domain.

Pro Tip — Use VLOOKUP between the two exported files above to connect the Source and Destination columns to the location of the final URL.

For example:

=VLOOKUP([@Destination],'response_codes_redirection_(3xx).csv'!$A$3:$F$50,6,

FALSE)

(Where 'response_codes_redirection_(3xx).csv' is the CSV file that contains the redirect links, and '50' is the number of lines in this file)

Using SEO Spider for internal linking

Internal linking can bring good ROI (return on investment), especially when you have a PageRank distribution strategy, keyword rankings and keywords containing anchor texts.

How to view Screaming Frog link position, path type and target

Screaming Frog SEO Spider v.13 records some new attributes for each link.

To see the position of each link in a crawl, such as navigation, page content, sidebar or footer, look at the crawl results. The classification is done using the path of each link (like XPath) and the known semantic substrings, which can be seen in the «Inlinks» and «Outlinks» tabs.

how to use screaming frog seo spider

If your site uses HTML5 semantic elements (or well-named non-semantic elements like div id="nav"), SEO Spider will be able to automatically detect the different parts of the web page and the links within them.

what is screaming frog used for

But not every website is built this way, so you can customize the classification of link positions in Config > Custom > Link positions. This allows you to substring the path of the link to classify it however you want.

For example, we have mobile menu links outside of the nav element that are defined as content links. This is not true, as they are additional mobile phone navigation.

how to find broken links in screaming frog

The class name «mobile-menu__dropdown» (which is in the link path as shown above) can be used to determine the correct position of the link using the «Links Positions» function. See below how to properly configure the Screaming Frog Seo Spider program. 

how to find broken links on screaming frog

These links will then be properly classified as navigation links.

how to use screaming frog to improve on page seo

This can help identify incoming links to a page only from the content, for example ignoring any links in the main navigation or footer for better analysis of internal links.

How to find out absolute and relative links 

Screaming Frog allows you to find out the path type of a link (absolute, relative, protocol, or root). It can be seen in links, outbound links, and all bulk exports.

 

This can help identify incoming links to a page only from the content, for example ignoring any links in the main navigation or footer for better analysis of internal links.

How to find out absolute and relative links 

Screaming Frog allows you to find out the path type of a link (absolute, relative, protocol, or root). It can be seen in links, outbound links, and all bulk exports.

what is screaming frog seo spider

This can help identify links that should be absolute, as there are integrity, security, and performance issues under some circumstances. You can sort and export only absolute links. 

Link Target Attribute

In addition, you can now look at the «target» attribute to detect links that use «_blank» to open in a new tab.

how to find orphan pages on a website screaming frog

This is useful when analyzing usability as well as performance and security, which brings us to the next feature.

3

Checking website content with Screaming Frog

How to find pages with minimal text content

Screaming Frog allows you to view pages with little or no content. Instructions: 

  1. After the algorithm has finished checking, open the «Internal» tab, sort by HTML and scroll right to the «Word Count» column. 
  2. Sort the «Word count» column from smallest to largest to find pages with the low text content. You can drag the «Word Count» column to the left to better understand which pages are related to a certain number. 
  3. Click «Export» in the «Internal» tab if you prefer to work with data in CSV format.

The above word count method determines the amount of actual text on a page. But there's no way to be specific — it's a product name or a keyword-optimized text block.

Pro Tips for E-commerce Sites.  To determine the word count of text blocks, use ImportXML2 to separate text blocks in any page list when counting characters from there. If xPath queries are not your forte, then the xPath Helper or Xpather Chrome extensions will do the solid job of determining xPath for you.

How to find duplicate content on a site

To find duplicate content, you need to go to the «Content» tab, which contains filters for «Near Duplicates» and «Exact Duplicates». 

how does screaming frog work

While there are no duplicate content penalties, having similar pages can cause crawling and indexing inefficiencies. Similar pages should be kept to a minimum, since high similarity is a sign of low-quality pages.

How to view content matching percentage

For «Near Duplicates», SEO Spider will show you the closest matching percentage as well as non-obvious duplicates for each URL. The «Exact Duplicates» filter uses the same algorithmic check to identify identical pages that were previously named «Duplicate» in the «URL» tab.

The new «Near Duplicates» tab uses the minhash algorithm, which allows you to set up a nearly identical threshold, which is set to 90% by default. This can be configured via Config > Content > Duplicates.

screaming frog how to find referring page

Semantic elements such as navigation and footer are automatically excluded from content checking, but you can refine it by excluding or including HTML elements, classes, and IDs. This can help focus the analysis on the main content area, avoiding well-known template text. It can also be used to provide more accurate word counts.

why is screaming frog showing deleted pages

Next to duplicates, analysis is required to complete after the scanning, and more details about duplicates can be seen on the new bottom tab «Duplicate details». This displays all identified nearly identical URLs and their matches.

how to build sitemap in screaming frog

By clicking on «Near Duplicate address» in the «Duplicate details» tab, you will see non-obvious duplicate content found among the pages.

screaming frog seo spider

The near-duplicate content threshold and content area used in the analysis can be updated after the crawl.

The Content tab also includes a «Low Content Pages» filter that identifies pages with less than 200 words using an improved count. This can be customized to your preference in Config > Spider > Preferences as there is no universal criteria for a minimum SEO word count.

Spelling and grammar checking

How to check spelling and grammar with Screaming Frog? The new «Content» tab contains filters for «Spelling Errors» and «Grammar Errors» and displays counters for each page crawled.

screaming frog seo spider

You can enable spelling and grammar check Config > Content > Spelling & Grammar

While this is a bit different from Screaming Frog's usual SEO-focused features, the main role of developers is to improve websites for users. Google's Search Quality Guidelines repeatedly describe spelling and grammatical errors as one of the characteristics of low-quality pages.

crawl site screaming frog

The right side of the details tab also displays the visual text from the page and the errors found.

screaming frog seo

The Spelling & Grammar tab in the right pane shows the 100 unique errors found and the number of URLs they affect. This can be useful for finding errors in templates, as well as building your dictionary or list of ignored words.

what is screaming frog

Screaming Frog's spelling and grammar checker automatically detects the language used on the page through the HTML language attribute, but also allows you to manually select it if needed.

In the Screaming Frog SEO Spider settings, you can update the spelling and grammar check to reflect changes in your vocabulary, ignore the list or grammar rules without rescanning URLs.

how to use screaming frog

You can export all data via the Bulk Export > Content menu.

how to use screaming frog seo spider

How to get a list of image links on a specific page

If you have already checked the whole site or a subfolder, simply select the page at the top of the window, click the «Image Info» tab at the bottom to see all the images found on the page. They will be listed in the «To» column.

Pro Tip — Right click on any entry at the bottom of the window to copy or open the URL.

Alternatively, you can also see the images on the page by only crawling that URL. Make sure Scan Depth is set to «1». Once the page has been crawled, click on the «Images» tab. There will be all the images found by the algorithm.

Alt check for images

Screaming Frog Seo Spider checks attributes for images: alt. How to do it: 

  1. Look at the settings and make sure «Check Images» is checked in Configuration > Spider. 
  2. After the check is complete, open the «Images» tab and sort by «Missing Alt Text» or «Alt text Over 100 Characters». 
  3. You can find image pages by going to the «Image Info» tab at the bottom of the window. These pages will be listed in the «From» column.
  4. If you're more comfortable working with CSV, open Bulk Export > All Images or Bulk Export > Images > Images Missing Alt Text Inlinks to see the full list of images. In the same place you will see their location, the alt text associated with them or problems with it.

what is screaming frog used for

How to find CSS files

In the Configuration > Spider menu, select «Crawl» and «Store» the CSS before checking. Once completed, sort the results by «CSS» in the «Internal» tab.

how to use screaming frog to improve on page seo

How to find JavaScript files

In the settings menu, select «Check JavaScript» before checking. Once completed, sort the results by «JavaScript» in the «Internal» tab.

how does screaming frog work

How to detect all jQuery plugins on a site and which pages they are used

Check the settings and make sure «Check JavaScript» is selected in the Configuration > Spider menu. After the check is complete, sort the results in the «Internal» tab by «JavaScript», and then search for «jquery». This will give you a list of plugins. If necessary, sort the list by «Address» for easier viewing. Then look for «InLinks» at the bottom of the window, or export to CSV to find the pages where the files are used. They will be displayed in the «From» column.

Also via Bulk Export > All Links menu you can export a CSV file and filter the «Destination» column to show URLs with jQuery only.

Pro tip — how to use Screaming Frog

Not all jQuery plugins are bad for SEO. If you see that the site is using jQuery, it's good to make sure that the content you want to index is included in the page source and is processed during page load, not after.

How to find linked internal PDF files

Once the scan is complete, sort the results by «PDF» in the «Internal» tab.

screaming frog how to find referring page

How to find pages with a share button on social media

To find pages that contain the «Share» button, you need to set up a custom filter before running the check. To do this, open Configuration > Custom. From there, enter any code snippet from the page source.

screaming frog seo spider

In the example above, we wanted to find pages that contain a Facebook Like button. To do this we created a filter for facebook.com/plugins/like.php

How to Find Pages That Use Frames

To find pages that use frames, set a custom filter for the iframe via Configuration > Custom before starting the test.

How to find pages with embedded video and audio content

To find pages with video or audio content, set a custom filter for the embed code snippet from Youtube or another media player that is used on the site.

screaming frog guide

4

Checking meta data: how to find duplicate titles and descriptions

How to find pages with long metadata 

To find long titles and descriptions, after the check is complete, open the «Page Titles» tab and sort by «Over 60 Characters» to see the long page titles. You can do the same in the «Meta Description» and «URL» tabs.

screaming frog seo spider metadata 

How to find duplicate titles and descriptions

To find duplicate title and description metadata, after checking is complete, open the «Page Titles» tab, then sort by «Duplicate». You can do the same in the «Meta Description» «URL» tabs.

screaming frog seo spider duplicate

How to find pages with meta-directives noindex, nofollow, etc.

Using Screaming Frog Seo Spider, you can find pages that are closed from indexing. After completing the check, go to the «Directives» tab. To see the type, scroll to the right and see which column is filled. You can also use the filter to find any of the following tags:

  • index;
  • noindex;
  • follow;
  • nofollow;
  • noarchive;
  • nosnippet;
  • noodp;
  • noydir;
  • noimageindex;
  • notranslate;
  • unavailable_after;
  • refresh.

This information can also be viewed in the right sidebar.

screaming frog seo spider sidebar

Is it possible to check the robots.txt file?  

Screaming Frog matches the robots.txt file by default. As a priority, it will follow the instructions made specifically for the Screaming Frog user agent. If there are no such instructions, the algorithm follows the instructions of Googlebot. If there are none, then the algorithm will execute the global instructions for all user agents. 

The algorithm will follow one set of instructions, so if there is a set of rules set only for Screaming Frog, it will only follow them. If you want to hide certain parts of the site from the algorithm, use the normal «Screaming Frog SEO Spider» robots.txt syntax. If you want to ignore robots.txt, just select this option from the settings menu: Configuration > robots.txt > Settings.

screaming frog seo spider robots.txt

How to find or check microdata on a site

To find pages with markup or microdata using Screaming Frog, use a custom filter. Go to Configuration > Custom > Search and enter the snippet you want. 

To find pages with markup, add the following code snippet to a custom filter: itemtype=http://schema.org

To find a specific kind of markup, specify the details. For example, using a custom filter will give you all pages marked up for ratings.

Speaking of Screaming Frog 11.0, the SEO algorithm also provides the ability to check, extract, and structured data directly from search results. Check any JSON-LD, Microdata or RDF structured data against Scema.org guidelines and Google specifications in real time during the process of checking. To access structured data checking tools, select the Configuration > Spider > Advanced > Structured Data option.

Now you will see the Structured Data tab in the main interface. It will allow you to switch between pages with structured data in which this data is omitted. They may also contain validation errors and warnings.

You can also bulk export structured data issues with Screaming Frog by visiting Reports > Structured Data > Validation Errors&Warnings.

screaming frog spider find or check microdata

5

Creating and validating a Sitemap in Screaming Frog

How to build sitemap in Screaming Frog (XML)

To create a sitemap.xml file with Screaming Frog Spider, after the site has been crawled, go to Sitemaps > Sitemap XML.

how to build sitemap in screaming frog

Once you open the Sitemap XML settings, you can include and exclude pages by response code, modification date, priority, change frequency, etc. Screaming Frog includes 2xx URLs by default, but double-checking is a good rule.

sitemap in screaming frog

Ideally, the sitemap should contain pages with status 200, the canonical version of each URL without any duplicate factors. Click OK to save your changes. The Sitemap XML will be downloaded to your device and will allow you to edit the naming convention at your discretion.

How to build an XML Sitemap by uploading a URL

You can also create an XML Sitemap by uploading URLs from an existing file or manually pasting them into Screaming Frog.

Change the mode from Spider to List in Mode and click Upload to select any of the options.

 build sitemap in screaming frog

Click the Start button and Screaming Frog will check the uploaded URLs. Once he does, you can do the above steps.

How to check an existing XML Sitemap

You can upload an existing XML Sitemap to check for errors or discrepancies.

Open the «Mode» menu in Screaming Frog and select «List». Then, click «Upload» at the top of the screen and select Download Sitemap or Download Sitemap Index, enter the sitemap URL and start checking. Once completed, you will be able to find redirects, 404 errors, duplicate URLs, etc.

How to check an existing XML Sitemap

Checking for missing pages in an XML Sitemap

You can set up validation options to examine and compare the URLs in the sitemap with the URLs from the search.

Go to Configuration > Spider in navigation and you will have several options for Sitemap XML - Auto discover XML sitemaps via robots.txt file or enter a link to Sitemap XML in the window.

Note: If your robots.txt file doesn't contain the correct target links for all the XML Sitemaps you want to scan, enter them manually.

Auto discover XML sitemaps

Once you've updated your sitemap validation settings, open «Crawl Analysis» and then click «Configure» and make sure the Sitemap button is checked. If you want to do a full site crawl first, then go back to Crawl Analysis and click Start.

sitemap validation settings

how to build sitemap in screaming frog

Upon completion of Crawl Analysis, you will be able to see all inconsistencies. For example, links that were identified during a full crawl, but missed in the sitemap.

6

Other technical problems that Screaming Frog will help to solve 

Why a particular section of a site isn't Indexed or ranked

Trying to figure out why certain pages aren't indexed? First, make sure they don't accidentally end up in robots.txt or are marked as noindex. We wrote above which Screaming Frog Seo Spider reports you need to watch.

Next, check that the algorithms can reach the pages by checking internal links. Pages without internal linking are referred to as Orphaned Pages. How to find orphan pages on a website using Screaming Frog? To identify such pages, follow these steps:

  1. go to Configuration > Spider. At the bottom there will be some options for the sitemap XML. Either auto-validate the sitemap via robots.txt or manually enter the sitemap XML link.
    Note that if robots.txt doesn't contain the correct target links to all the XML sitemaps you want to scan, you'll need to enter them manually;
  2. go to Configuration > API Access > Google Analytics. Using the API, you can get and view analytics data for a specific account. To find orphan pages from organic search, make sure to segment by «Organic Traffic»;what is screaming frog used for
  3. you can also go to the General > Crawl New URLs Discovered in Google Analytics tab if you want them to be included in the full site check. If this setting is not enabled, you will only be able to view new URLs from Google Analytics in the Orphan Pages report;how to use screaming frog seo spider
  4. go to Configuration > API Access > Google Search Console. Using the API, you can get and view Google Search Console analytics data for a specific account. To find orphan pages, you can search for URLs getting clicks and impressions that weren't included in the crawl. Tip: You can also go to General > Crawl New URLs Discovered in Google Search Console if you want them to be included in the full site check. If this setting is not enabled, you will only be able to view new URLs from Google Search Console in the Orphan Pages report;
  5. crawl the whole site. Once completed, go to Crawl Analysis > Start and wait for it to complete;
  6. see the orphan URLs in each tab, or bulk export the URLs by going to Reports > Orphan Pages.

what does screaming frog do

If you do not have access to Google Analytics or Search Console, you can export the list of internal URLs as a CSV file using the «HTML» filter in the «Internal» tab.

Open the CSV file and on the second page paste a list of URLs that aren't indexed or don't rank very well. Use VLOOKUP to see the URLs found during the crawl on the second page.

How to check if the site migration was successful

You can use Screaming Frog to check if old URLs were redirected using List mode to check code statuses. If the old links return a 404 error, then you will understand which URLs should be redirected.

How to find pages with a slow loading speed

​​Screaming Frog will help you find pages with a slow loading speed. After the scan is complete, open the «Response Code» tab and sort the «Response Time» column in descending order. This way you can identify pages with the slow loading speed.

how to use screaming frog seo spider

How to find malicious software or spam on a site

First you need to identify traces of malicious software or spam. Next, go to Custom > Search in the settings menu and enter the snippet you are looking for.

How to find malicious software or spam on a site

You can enter up to 10 fragments per scan. Then click OK and continue scanning the site or list of pages.

After the scan is complete, select the «Custom» tab at the top of the window to see all pages that contain the specified fragment. If you have entered more than one custom filter, you can see each one by changing it in the results.

7

PPC verification and analytics

How to validate a PPC URL list

Save the list as a .csv or .txt file. Change the «Mode» setting to «List».

Then select a file to upload. You can click Start, or paste the list into Screaming Frog manually. Look at the status code of each page in the «Internal» tab.

8

Cleaning up information in Seo Frog Spider

How to clean up metadata for a page list

So you have a bunch of URLs, but you need more information about them? First of all, set the mode to «List». Upload the list of URLs (.csv or .txt file). Once verified, you can see code statuses, outbound links, word counts, and meta data for each page on your list.

How to clean up the site from pages that contain certain information

Go to «Settings», select Extraction or Search in the «Custom» menu and enter the desired fragment.

You can enter up to 10 different snippets per search. Click OK and proceed to crawl the site or list of pages. In the example below, we wanted to find all pages that contain «Please Call» in the pricing section, so we found and copied the HTML code from the page's source.

how to use screaming frog to improve on page seo

Upon completion of the search, select the «Custom» tab at the top of the window and see all the pages with the required fragment. If you have entered more than one custom filter, you can view each one by changing the result filter.

Pro Tip — If you're pulling data from a client's site, you can save time by asking the client to pull the data directly from the database. This method is suitable for sites that you do not have direct access to.

9

Renaming URLs in program settings

How to find and replace session IDs or other parameters on scanned pages

To identify URLs with session IDs or other parameters, simply scan your site with default settings. Upon completion of the check, go to the «URL" tab and sort by «Parameters» to see all links with parameters.

To remove parameters from displaying in validated URLs, go to Configuration > URL Rewriting. Then in the «Remove Parameters» tab, click «Add» to add the parameters you want to remove from the URLs and click OK. Restart the scan with these overwrite settings.

what is screaming frog

How to rename verified URLs

If you need to rename URLs after crawling a site with Screaming Frog Seo Spider, go to Configuration > URL Rewriting to rewrite any verified URL. Then, in the «Regex Replace» tab, click «Add» to add the Reg Ex for what you want to replace.

how to use screaming frog

Once you have added all the desired rules, you can test them in the «Test» tab. To do this, enter a test URL in «URL before rewriting». «URL after rewriting» will be updated automatically according to your rules.

 

Screaming Frog: URL after rewriting

If you want to set a rule that all URLs return in lowercase, set «Lowercase discovered URLs» in the «Options» tab. This will remove any duplication on header URLs in the check.

what is screaming frog

Don't forget to restart the scan with these overwrite settings.

10

Competitor’s keyword analysis

How to find out my competitors' most important pages

In general, competitors will try to spread link popularity and drive traffic to the most valuable pages using internal linking. Any competitors with a SEO mindset are more likely to link to important pages from the company's blog. Find valuable pages of your competitors by scanning their site. Then sort the «Inlinks» column in the «Internal» tab in descending order to see the pages with the most internal links.

how to use screaming frog

To view pages linked from your competitor's blog, deselect «Check links outside folder» in Configuration > Spider and scan the blog folder. Then, in the «External» tab, sort the results using the search by primary domain URL. Scroll to the right and sort the list in the «Inlinks» column to see the most linked pages.

what does screaming frog do

Pro Tip — Drag a column to the left or right to improve the display of data.

What anchors do my competitors use for internal links?

Screaming Frog Seo Spider allows you to analyze anchors on your site or competitors' anchors. 

Go to Bulk Export > All Anchor Text to export a CSV file that contains all the anchor texts on the site, where they are used and what they link to.

how to use screaming frog seo spider

11

How to find a site's outbound links and use it for link building

How to get a list of promising link locations

If you've deleted or created a list of URLs you need to check, you can download and scan them in «List» mode to get more information about the pages. Check the status codes in the «Response Codes» tab, when the scan is complete. View the outgoing links, their types, anchor texts and nofollow directives in the «Outlinks» tab in the lower window. This will tell you which sites these pages link to. To view the Outlinks tab, make sure the URL you are interested in is selected in the top window.

Of course, with a custom filter, you'll want to check if these pages are linking to you or not.

 screaming frog instructions

You can also export the full list of outlinks by going to Bulk export > All outlinks. This will not only give you links to external sites, but it will also show you all the internal links on the individual pages of your list.

screaming frog leading

How to find broken links for outreach opportunities

So, have you found the site you want to get a link from? Use Screaming Frog to find broken links on the page or site, then reach out to the site owner, suggesting your resource as a replacement for the broken link where possible. Well, or just offer a broken link as a gesture of good faith :)

How to check backlinks and see the anchor text

Load your list of backlinks and run the algorithm in Mode > List mode. Then export the full list of outbound links by going to Bulk Export > All Outlinks. This will give you the URLs and anchor/alt text for all links on those pages. You can then use the filter on the «Destination» column in the CSV to determine if your site is being linked and what anchor/alt text is included.

How to make sure backlinks have been removed

Set a custom filter containing the root domain URL by going to Configuration > Custom > Search. Then load the list of backlinks and run the algorithm in Mode > List mode. Once the scan is complete, click the «Custom» tab to see all the pages that still link to you.

Bonus Round

Did you know that when you right click on any URL in the top results window, you can do any of the following?

  • copy or open URL;
  • re-crawl or remove the URL;
  • export URL information, inbound links, backlinks, or image information for this page;
  • check page indexing in Google, Bing and Yahoo;
  • check backlinks to the page in Majestic, OSE, Ahrefs and Blekko;
  • look at the cached version and cache date of the page;
  • view old versions of the page;
  • check the HTML code of the page;
  • open the robots.txt file for the domain where the page is located;
  • find other domains on the same IP. 

screeming frog directions

12

How to edit metadata

SERP mode allows you to view SERP snippets on the device to visualize metadata in search results.

  1. Upload URLs, titles and meta descriptions to Screaming Frog using a CSV or Excel document.
    Note that if you have already crawled your site, you can export URLs by selecting Reports > SERP Summary. This will easily format the URLs and meta you want to download and edit.
  2. Go to Mode > SERP > Upload File.
  3. Edit the metadata in Screaming Frog.screaming frog seo spider regulation
  4. Make a bulk export of metadata to send directly to developers.

How to analyze a JavaScript website

JavaScript frameworks such as Angular, React, etc. are increasingly being used to build websites. Google recommends using a rendering solution as Googlebot is still trying to scan the JavaScript content. If you have defined a site using JavaScript and want to crawl, follow the instructions below:

  1. go to Configuration > Spider > Rendering > JavaScript;screaming frog seo spider instructions
  2. change the render settings depending on what you are searching for. You can adjust the timeout, window size (mobile, tablet, desktop);
  3. click OK and start scanning.

At the bottom of the navigation, click the «Rendered Page» tab to see how the page is displayed. If it doesn't display properly, check for blocked resources or increase the timeout limit in the configuration settings. If none of the options solve it, a more serious problem will arise.

You can view and bulk export any blocked resources that may affect your site's crawling and rendering by going to Bulk Export > Response Codes.

screaming frog seo spider precept

We hope this guide will help you understand the Screaming Frog Seo Spider settings. In this guide, we have collected the main features of Screaming Frog: how to do a technical audit, how to find duplicate metadata, how to check outgoing links, how to check broken links and redirects, how to use Screaming Frog to simplify link binding and other nuances of working with the program.

Did you know about all the features of Screaming Frog Seo Spider?