A Google Search Console sitemap guides search engines through a site. It lists URLs, metadata and priorities. It helps sites get crawled and indexed faster. This guide shows how to build, submit and maintain a sitemap in Search Console.
Understand sitemap basics
A sitemap is a file that maps a site’s structure. It tells crawlers which pages to fetch. Key sitemap benefits include faster discovery of new or updated content.
Define a sitemap
A sitemap provides details about site URLs. It can include:
lastmod: date a page last updatedchangefreq: suggested crawl intervalpriority: relative importance of each page- image and video metadata
Types of sitemaps
Search engines accept XML, RSS, mRSS and Atom formats. XML sitemaps work best. HTML sitemaps help visitors navigate. A sitemap index groups multiple sitemaps for large sites.
Generate a sitemap
A sitemap tool or plugin can automate file creation. Many content management systems include built-in sitemap features.
Handle large sites
Google limits a single sitemap to 50 000 URLs or 50 MB uncompressed. For larger sites:
- split sitemaps by section or date
- use a sitemap index file
- reference each sitemap in the index
Encode and host file
Sitemaps must use UTF-8 encoding. Hosting at the site root impacts all site files. Use fully qualified URLs that include protocol and domain.
Submit and verify sitemaps
Site owners should confirm their property in Search Console. See google search console setup.
Submit via Search Console
Site owners submit a sitemap using the Sitemaps report. They follow these steps:
- Site owners navigate to the Sitemaps report in Search Console.
- They enter the path to the sitemap (for example, sitemap.xml).
- They click Submit.
- They review status and discovered URL count.
Add sitemap in robots.txt
Site owners can hint crawlers via robots.txt. They take these steps:
- They open the robots.txt file
- They add
Sitemap: https://example.com/sitemap.xml - They save and upload changes
Monitor and troubleshoot
Regular checks keep sitemaps error-free. Site owners should review status after each major update.
Check sitemap status
The Sitemaps report shows last crawl date and URL count. A green status means Google read the file.
Diagnose sitemap errors
Search Console flags fetch or parse issues. It retries fetching for several days before stopping.
Common errors include:
- fetch error: network or permission issue
- parse error: syntax or encoding mistake
- URL error: blocked or nonexistent page
Resolve sitemap issues
Site owners fix reported errors then resubmit sitemaps. They ensure the file uses valid XML, UTF-8 encoding and correct URLs.
Optimize for indexing
Rich metadata boosts crawl efficiency and indexing speed.
Use metadata tags
XML sitemaps should include:
<lastmod>to show update dates<changefreq>to suggest crawl intervals<priority>to rank page importance
Include multimedia content
Sitemaps can list image and video details:
<image:image>with image URLs and titles<video:video>with video titles and durations
Manage alternate URLs
Sitemaps support language and device variants:
- use
<xhtml:link>for hreflang versions - include separate mobile and desktop URLs when needed
Maintain a sitemap
A well-maintained sitemap reflects live site content.
Schedule regular updates
Site owners regenerate sitemaps after publishing new content. Automation via cron jobs or CMS triggers works best.
Remove outdated URLs
Deleting or 301-redirecting removed pages prevents crawlers from fetching dead links.
Resubmit after changes
Resubmission in Search Console speeds recrawl of updated content.
Explore further resources
- Site owners find setup steps in google search console setup
- Site owners review usage tips in how to use google search console
- Site owners track URLs with google search console performance report
- Site owners explore crawl insights in google search console indexing