Read time: 12 minutes
How often do we hear that the big things are the little things? In the case of launching a new website or a website redesign, spoiler alert, it all comes down to the little things. You can have the most beautifully designed website, a perfectly architected site structure and even best practice search engine optimization tactics, but if you miss the all-important technical elements, there’s a good chance that your site will end up in the boonies when it comes to search engine results. And what good is your beautiful new website, with perfectly crafted content if it doesn’t get any visitors?
What is a Sitemap and do I need one?
Simply put, a sitemap is a file that can be sent to search engines, listing the pages on your website. Sitemaps are effectively used to communicate with search engines to let them know that you have web pages that they ought to crawl and subsequently index, as the content is valuable and worthy of visitors.
There are two types of sitemaps, one is a visual sitemap intended as a navigation tool for visitors to your site and the other is an XML sitemap that’s geared towards SEO. Today we’ll be focusing on the latter which is a file that lists all of the pages on your website and is read by search engine crawlers (such as Googlebot and Facebot) to enable them to more intelligently crawl your website. XML sitemaps provide search engine crawlers with an insight into the information that’s published on your website and when it was last updated.
The queen of all search engines; Google, have confirmed that XML sitemaps can improve the crawling of your site. Sitemaps are helpful as they provide search engines with an indication that your content is good quality and worthy of indexing, but a sitemap alone won’t completely transform your SEO. However, there’s no doubt that by not having a sitemap, you are taking a risk to your website traffic and it’s something that’s overlooked regularly. Time and time again when we conduct SEO audits, we’re often asked the golden question, ’what is a sitemap?’. Web designers are great at website design but SEO tools such as sitemaps aren’t typically at the forefront of their considerations when creating an alluring website, so it’s important to have an SEO expert on-hand as well.
Should I Include Every Webpage in the XML Sitemap?
The short answer is no. You probably don’t want all of the pages on your website indexed so it’s ineffective to indicate that all webpages contain quality content that ought to be indexed, when in reality this doesn’t apply to every page on your website. For example, is the ‘Thank You for Contacting Us’ page that appears when a visitor submits a ‘Contact us’ form, high quality content that’s worthy of indexing? Probably not.
It’s common practice to have a number of pages on your website like this that fall into the category of ‘utility’ pages. This is totally expected but it’s fundamental to keep such utility pages away from your Sitemap because if they’re included it can indicate to Google that you don’t understand the meaning of quality content in the context of search engine results pages (SERPs). On the other hand, if your website contains 250 pages, 100 of which are utility pages which are omitted from your XML sitemap and 150 of which are included in your sitemap as they contain content that’s relevant for indexing and display on SERPs, then you’re sending a clear message to Google and increase your chances of ranking for quality content.
Consistency is vital when it comes to communicating with Google. So if you include a page as “noindex” in your robots.txt file then it shouldn’t be included in your XML sitemap. Just because you’ve omitted some pages from your sitemap doesn’t mean they won’t appear on search engine results pages (SERPs). For pages that you wish to hide from (SERPs) you’ll have to create a robots.txt file, as explained in detail below.
How do I create a Sitemap?
The good news is there are lots of tools available to help with this, most of which are free. Yoast is available as a plug-in for WordPress and can be used as an effective tool to easily create an XML sitemap. Once Yoast has been installed, simply enable the plugin by clicking on the features tab and then enable ‘Advanced settings pages’. Once Yoast has been enabled on WordPress, a tab for ‘XML Sitemaps’ will appear on the sidebar. Simply set this to ‘enabled’ and you’re off to the races, your sitemap has been created! Once you’ve created your sitemap, it needs to be submitted to Google (through Search Console) and Bing (through Webmaster Tools) it’s that easy.
What is a Robots.txt file?
A robots.txt file is a useful file created by webmasters to instruct web robots, typically search engine crawlers, how to crawl pages on their websites. A robots.txt file is part of the Robot Exclusion Protocol (REP), which are web standards which manage how bots crawl websites and index content to serve on search engine results pages (SERPs).
Why Block a Search Engine from Indexing a Webpage?
There are four main reasons why it makes sense to block search engine crawlers from indexing a webpage using the robots.txt file. First, if you have webpages with duplicate content then it makes sense to block duplicate pages from being indexed as duplicate content can negatively effect your SEO. Second, some webpages are presented when an action is completed. For example, a thank you webpage is often displayed after a form is completed. In this instance, you probably don’t want a search engine to index the thank you page, as it only makes sense for it to be displayed after an action is taken. Third, perhaps you want to use a robots.txt file to stop a search engine from indexing a page to maintain privacy, but bear in mind that not all bots adhere to the stipulations of a robots.txt file. Lastly, robots.txt files can help keep your bandwidth from being used up by search engine robots indexing your images, which can impact page load speed, again something that can negatively impact your SEO.
How to Create a Robots.txt file
Thankfully, Google makes this process pretty straightforward, all you have to do is set up a free Google Search Console account and create a robots.txt file with the following steps:
- Click on ‘Generate robots.Txt’.
- Under the ‘Action’ tab, select ‘Block’.
- Under the ‘User Agent’ tab, you can choose to block ‘all robots’ from crawling the specified pages or just block particular robots e.g. Google-mobile.
- Add relative links (e.g /Thank-you rather than www.websitedomain.com/thank-you’) to the pages you wish to block under ‘Directories and Files’.
- Click on ‘Add rule’ to create a rule for specified pages.
- ‘Download your robots.txt file’ to generate the robots.txt file as specified.
- Once you’ve successfully created a robots.txt file, it needs be uploaded to the root of your domain as a text file . The file should be named robots.txt and /robots.txt should be the URL for your robots.txt file.
Don’t forget to update your robots.txt file as you update your website with pages that you don’t want to be indexed by search engines to maintain the privacy of your website while getting the best out of your search engine optimization activities. Below is an example of the New York Times’ robots.txt file:
Essential Robots.txt file Commands
User-agent: This refers to the web crawling robot to that is being addressed for a specific command. It’s important to note that some search engines have multiple user agents e.g. ‘Googlebot’ refers to Google’s crawler for organic search and ‘Google-bot image’ for image search. Most user agents from the same search engine follow the same rules so it’s not necessary to specify for each of their user agents, but it’s handy to have the ability to do so when appropriate.
Allow: This command only works with Googlebot (Googles user agent) and encourages the bot to crawl a specified webpage, even if the parent page or subfolder has been blocked. A separate command is required for each webpage you wish to allow.
Disallow: This command tells the specified user agent not to craw the specified page. A separate command is required for each webpage you wish to disallow.
Crawl-delay: This directs a user agent to wait for a number of seconds before crawling a page. This command is not followed by Googlebot, however a crawl-delay can be set in Google Search Console as required.
Sitemap: This is used to specify the location of any XML sitemaps that are associated with a particular webpage. This command is supported by Google, Bing and Yahoo.
No Follow: This command directs a user agent not to follow the outbound links included on the page.
Robots.txt Tips for Success:
- Robots.txt files are publicly accessible, simply add ‘/robots.txt’ to the end of a website domain to view its robots.txt file, so don’t use it to hide sensitive information, as anyone can see which pages the webmaster wants to be crawled or not crawled.
- A robots.txt file must be placed in the websites top level directory (typically in the root domain or homepage) to be successfully found and interpreted by user agents, otherwise it’s likely to be unfound and therefore disregarded.
- A robots.txt file is case sensitive and the file must be called ‘robots.txt’ and no other variation.
- A subdomain on a root domain uses separate robots.txt files e.g. blog.domain.com and domain.com, so it’s important to create two separate robots.txt files for each.
- Best practice suggests to add the sitemap location at the bottom of a robots.txt file.
- Make sure you’re not blocking any sections of your site that you do want crawled as this can have a detrimental impact on your SEO practices.
- Some user agent robots disregard the robots.txt file, particularly malicious crawlers that skim information.
- Search engines typically cache robots.txt files and update them once a day so if you make changes to your robots.txt file and and want them to be incorporated sooner than that, simply submit the new robots.txt file URL through Google Search Console.
Search Console – Waste of Time or Time Well Wasted?
If you’re familiar with Google Webmaster Tools, chances are you know about Google Search Console, its name since 2015, to encompass the wide group of users including designers, SEO experts and marketers. In a nutshell, Google Search Console is a free service that provides insights to help you measure your website performance, flag potential issues and control how Google views your website. Below, you can find out how to set up Google Search Console but before you can reap the benefits, you’ll need to first add and verify your authority to access the website. Given the confidential nature of the information supplied, Google needs to ensure that you’re the site owner or webmaster before sharing the delights of Google Search Console with you.
How to Set up Google Search Console
1. Log into your Google Search Console account, enter the website URL and click ‘Add property’ to get started.
2. Next, you’ll be presented with the four options below to complete Google Search Console verification – HTML file, HTML tag, Google Analytics, Google Tag Manager or Domain name provider.
3. You’ll be presented with a confirmation once the website is verified. After verification, it’s important to not remove the HTML file upload, HTML tag or Google Tag Manager or your site will become unverified.
What’s the Difference between Google Search Console and Google Analytics?
While both free tools provide information that relates to your website, the information provided is different. Google Analytics focuses on information around visitors to your site such as how they arrive at your website, time spent on each page, geographic region etc. On the other hand, Google Search Console is focused on highlighting issues relating to your site such as broken links, malware, keywords that drive traffic to your site.
It’s important to note that if you look at the same report in Google Analytics and Google Search Console separately, the results may be different as the tools look at the information from different perspectives. It’s worthwhile linking both accounts as you will benefit from additional reporting capabilities. Simply log into Google Search Console and once you’ve verified your website based on the steps above, and select ‘Google Analytics Properly’ from within the settings icon in the top right hand corner.
Understanding Site Errors in Google Search Console
Gain an Insight into your SEO Performance with Google Search Console
Search Console can show you how often your site appears in Google search queries so that you can monitor your SEO performance. Visit the ‘Search Analytics’ tab in the ‘Search Traffic’ section of the sidebar. This gives you an insight into your search traffic over time and where your traffic is coming from.
This powerful tool provides the information you need to enhance your SEO by understanding more about your traffic source, visitor device type and which pages have the highest click-through-rates. It’s helpful to understand more about your mobile traffic so that you can ensure an effective user-experience for pages with significant mobile traffic.
How to Re-Index an Updated Page or Entire Website?
You can manage and update your sitemap and robots.txt file within Google Search Console. This is particularly handy if you make changes to your website that you want Google to index fast.
- Select the ‘Crawl’ tab and then click on ‘Fetch as Google’ and you’ll be presented with a screen to enter the URL to the page that you want re-indexed. If you made changes right across your website or on your homepage, then you’ll want to leave the URL box blank so that your entire site is crawled.
- Click ‘Fetch and Render” which can take a few minutes.
- Once your site has been re-indexed, scroll to the bottom and click on the ‘Submit to Index’ button. Choose ‘Crawl Only this URL’ for a single page and select ‘Crawl this URL and Direct Links’ if you want to index the entire site.
- Once the indexing is complete, you’re all set! The changes will appear on Google within the next few days.
Given the powerful insights and tools provided by Google Search Console, it should be very clear that this impressive free tool is an absolute must-have for monitoring and improving your website performance!
I hope you found the above overview of sitemaps, robots.txt files and Google Search Console useful and informative for getting the technical aspects of your website in check. While sometimes overlooked, these technical components can make all the difference to achieving a successful SEO strategy, giving you the ability to take your website from good to great!
Please note: We do not actively update blog posts with new release notes from these critical SEO tools. This post should not be used as the sole source of information on how to use Sitemaps, Robot.txt and GSC.
We’d love to hear more about your experience with sitemaps, robots.txt files and Search Console so let us know what you find most useful about these tools in the comments below!
A graduate of the Management of Technology MBA program from Simon Fraser University (SFU) and a former Board of Director for the SFU Alumni board, Erica finds balance through academic challenges as well.
As a self-diagnosed digital marketing junky, she spends most of her days studying search engine algorithm updates, A/B testing her own confirmation biases and building digital marketing strategies for various B2B clients.
When unplugged from the Digital World, Erica spends her time brewing beer, ultramarathon running and exploring the vast north of the 49th parallel with her husband, kids and vizsla.
Latest posts by Erica Hakonson (see all)
- The Critical Importance of Sitemaps, Robots.txt and Google Search Console (GSC) for SEO Success - August 14, 2019
- 7 Deadly Sins of Email Marketing List Management & Database Management - June 19, 2019
- Should You be Using Google’s Marketing Platform? - January 31, 2019