Sitemap Checker & XML Sitemap Validator
Instantly check whether your website's sitemap.xml exists, is publicly accessible, and passes key validation checks — including XML structure, URL count, sitemap index detection, and common errors. Free, no signup required.
What is a sitemap.xml?
The XML file that tells search engines — and AI crawlers — what pages your site has and when they were last updated.
A sitemap.xml is a structured XML file placed at the root of your domain (e.g. https://example.com/sitemap.xml) that lists the pages on your website. It is part of the Sitemap Protocol, which is supported by Google, Bing, Yahoo, and other major search engines.
While search engine crawlers will discover most pages through internal links, a sitemap provides a reliable fallback — especially for large sites, new content, or pages buried deep in a site architecture. It is particularly useful for pages that aren't well-linked internally.
Sitemaps can also be used for image sitemaps, video sitemaps, and news sitemaps — each using specialised extensions of the same XML format to surface media content to search engines.
Sitemap Index vs Standard Sitemap
<urlset> (Standard)Lists individual page URLs directly. Max 50,000 URLs per file.
Best for: Sites up to ~50,000 pages
<sitemapindex> (Index)References multiple child sitemap files. Used for large or segmented sites.
Best for: Large or multi-section sites
Why a Valid Sitemap Matters for SEO
Faster Crawl Discovery
Search engines don't have to rely solely on following links to find your pages. A sitemap acts as a direct map — new and updated pages get discovered and potentially indexed faster.
Large Site Management
For sites with thousands of pages, a sitemap index makes it easy to organise URLs into logical groups (blog, products, authors) and ensure complete coverage without hitting the 50,000 URL per file limit.
Freshness Signals
The <lastmod> tag in your sitemap tells crawlers when a page was last updated. Accurate lastmod dates help prioritise crawl frequency for content that changes regularly.
An honest perspective on sitemaps and rankings
A sitemap alone does not improve your Google rankings. It helps search engines find your pages, but whether those pages rank depends on content quality, authority, and relevance. Think of a sitemap as removing a technical obstacle — not as a ranking lever. A site with excellent content but no sitemap will almost always outrank a poorly-written site with a perfect sitemap.
Sitemap Best Practices for Technical SEO
What separates a good sitemap from a problematic one — and common mistakes to avoid.
Only include indexable pages
Your sitemap should list only pages you want indexed. Exclude pages with noindex directives, pagination variants, thin content pages, 404s, redirects, or canonicalized duplicates.
Use accurate <lastmod> dates
The <lastmod> tag should reflect when the page's content last substantively changed — not simply when the CMS ran a batch update. Inaccurate lastmod dates can cause crawlers to deprioritise your sitemap data.
Stay within the 50,000 URL limit
Each sitemap file is limited to 50,000 URLs and 50MB. If your site exceeds this, use a sitemap index file that references multiple individual sitemap files — each within the limit.
Reference your sitemap in robots.txt
Add a Sitemap directive to your robots.txt file: "Sitemap: https://yourdomain.com/sitemap.xml". This ensures any crawler that reads your robots.txt can discover your sitemap, even without a Search Console submission.
Submit to Search Console and Bing Webmaster Tools
Submit your sitemap directly in Google Search Console and Bing Webmaster Tools. This gives you visibility into how many URLs were submitted vs indexed, and any errors the crawlers encountered.
✗ Common Mistakes
- —Including noindex or redirected pages
- —Using relative URLs instead of absolute ones
- —Exceeding the 50,000 URL per file limit
- —Serving sitemap with wrong Content-Type
- —Returning a non-200 HTTP status
- —Using inaccurate or auto-generated <lastmod> dates
✓ Best Practices
- —List only pages you want indexed
- —Use full absolute HTTPS URLs
- —Include accurate <lastmod> dates in ISO 8601
- —Use sitemap index for sites over 50,000 pages
- —Reference sitemap in robots.txt
- —Submit to Search Console and monitor coverage
Example sitemap.xml
<?xml version="1.0" encoding="UTF-8"?>
<urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9">
<url>
<loc>https://example.com/</loc>
<lastmod>2026-03-01</lastmod>
</url>
<url>
<loc>https://example.com/about</loc>
<lastmod>2026-01-15</lastmod>
</url>
<url>
<loc>https://example.com/blog/my-post</loc>
<lastmod>2026-02-20</lastmod>
</url>
</urlset>⚠ Note: Google ignores <changefreq> and <priority> values. Use <lastmod> to signal content freshness — it's the only date field that influences crawl prioritisation.
How This Sitemap Checker Works
What we check — and what each result means.
HTTP 200 response
We fetch /sitemap.xml and confirm the server returns HTTP 200. Non-200 responses (404, 403, 5xx) are flagged.
Content-Type header
We verify the sitemap is served with an XML-appropriate Content-Type (application/xml or text/xml).
XML structure
We detect whether the file has a valid <urlset> or <sitemapindex> root element. Malformed XML is flagged.
Sitemap type detection
We identify whether this is a standard URL sitemap or a sitemap index file, and display the appropriate metrics.
URL / child sitemap count
We count <url> entries in standard sitemaps and <sitemap> entries in index files. Empty sitemaps are flagged as errors.
URL validity
We sample <loc> values and flag relative URLs, empty <loc> tags, or malformed URLs that violate the Sitemap Protocol.
Duplicate URL detection
We check for duplicate <loc> entries within the sampled portion of the sitemap.
<lastmod> presence
We check whether the sitemap includes <lastmod> dates, which help search engines understand content freshness.
URL limit warning
We flag standard sitemaps approaching or exceeding the 50,000 URL per file limit.
Understanding Results
This check meets the Sitemap Protocol or best practice standards. No action needed.
A recommended improvement. Your sitemap is still functional, but this is worth addressing.
A critical issue that should be fixed. Search engine crawlers may not process this sitemap correctly.
Frequently Asked Questions About Sitemaps
What is a sitemap.xml?+
Why is having a sitemap important for SEO?+
What is a sitemap index?+
What is the URL limit for a sitemap?+
How do I submit my sitemap to Google?+
What should a sitemap.xml include?+
Should I include noindex pages in my sitemap?+
What does this sitemap checker actually test?+
My sitemap returns HTTP 200 but search engines are not indexing all my pages — why?+
Does a sitemap improve my Google rankings?+
Where should a sitemap be located?+
Check Another Domain
Run the sitemap validator on any website — a competitor, a client site, or your own domain after making improvements.
Want deeper AI visibility insights? GEOflux tracks your brand across ChatGPT, Gemini, and Perplexity.
Start a free trial →