Start here
Before You Fix It: What This Check Means
Sitemaps are crawl-discovery hints that help engines find and revisit canonical URLs. In plain terms, this checks whether search engines have a sitemap they can actually discover and use. Scavo tries sitemap URLs in two ways.
Why this matters in practice: incorrect signals here can dilute indexing clarity and search traffic quality.
How to use this result: treat this as directional evidence, not final truth. Search indexing outcomes depend on crawler recrawl cadence and ranking systems outside your direct control. First, confirm the issue in live output: verify raw HTML output and crawler-facing validators Then ship one controlled change: Serve sitemap XML from a stable, publicly accessible URL. Finally, re-scan the same URL to confirm the result improves.
Background sources
TL;DR: Your sitemap.xml is missing or broken, forcing Google to discover pages through crawling alone and potentially missing new content.
Sitemaps help search engines use crawl budget more effectively by presenting a clear page hierarchy — especially critical for larger sites (Google Search Central). Without one, Google relies on link discovery alone, which means new pages, updated content, and deep pages may take weeks to get indexed. Note: Google ignores priority and changefreq values, so focus on accuracy and completeness.
What Scavo checks (plain English)
Scavo tries sitemap URLs in two ways:
- Reads
Sitemap:directives from yourrobots.txt(if available). - Probes common fallback endpoints:
/sitemap.xml/sitemap-index.xml/sitemap_index.xml/sitemap/sitemap.xml/sitemap/sitemap-index.xml/sitemap/sitemap_index.xml
Exact logic:
Pass: at least one candidate returns HTTP200and looks like sitemap XML.Warning: candidate returns200but body looks non-XML/non-sitemap.Warning: no valid sitemap found at tested candidates.
Scavo checks structural sitemap signals (status/content shape), not just URL existence.
How Scavo scores this check
Scavo assigns one result state for this check on the tested page:
- Pass: baseline signals for this check were found.
- Warning: partial coverage or risk signals were found and should be reviewed.
- Fail: required signals were missing or risky behavior was confirmed.
- Info: Scavo could not gather enough reliable evidence on this run to score pass/fail confidently.
In your scan report, this appears under What failed / What needs attention / What is working for sitemap, followed by Recommended next steps and Technical evidence (for developers) when needed.
- Scan key:
sitemap - Category:
SEO
Why fixing this matters
Sitemaps support faster discovery and refresh of your key URLs, especially on larger or frequently updated sites.
Without a valid sitemap path, crawlers can still discover pages via links, but discovery is slower and less predictable.
Common reasons this check warns
- Sitemap endpoint moved but robots.txt was not updated.
- Endpoint returns HTML (app shell/login page) with HTTP 200.
- CDN/proxy route rewrites
/sitemap.xmlincorrectly. - Sitemap generation job failed silently.
If you are not technical
- Ask who owns sitemap generation (CMS plugin, framework job, custom service).
- Confirm your live sitemap URL and request proof it opens as XML.
- Ensure robots.txt references the live sitemap URL.
- Re-run scan after deployment.
Technical handoff message
Copy and share this with your developer.
Scavo flagged XML Sitemap (sitemap). Please ensure at least one sitemap endpoint returns HTTP 200 with valid sitemap XML and that robots.txt references the live sitemap URL. Share endpoint proof and re-run the scan.If you are technical
- Serve sitemap XML from a stable, publicly accessible URL.
- Keep robots.txt
Sitemap:directive in sync. - Return correct content type and XML body (
<urlset>or<sitemapindex>). - Avoid routing fallback that serves HTML at sitemap paths.
- Rebuild sitemap when major IA changes ship.
Example robots.txt line
Sitemap: https://www.example.com/sitemap.xml
How to verify
curl -I https://www.example.com/sitemap.xmlreturns200.- Body begins with XML sitemap structure.
- Google Search Console Sitemaps report accepts the URL.
- Re-run Scavo and confirm
Pass.
What this scan cannot confirm
- It does not fully validate every URL listed inside sitemap files.
- It does not confirm
lastmodaccuracy or priority strategy. - It only checks discovered/common endpoints for this run.
Owner checklist
- [ ] Assign owner for sitemap generation and publishing.
- [ ] Add alerting when sitemap job/output fails.
- [ ] Keep robots.txt sitemap directives version controlled.
- [ ] Re-validate sitemap after route or domain migrations.
FAQ
Can we rank without a sitemap?
Yes, but crawlers rely more on internal linking discovery. Sitemaps improve reliability and freshness signals.
Why does Scavo warn when a sitemap URL returns 200?
Because the check also validates that the response looks like sitemap XML. A 200 HTML response at /sitemap.xml is still a broken sitemap setup.
Should we have one sitemap or many?
Either can work. Large sites often use a sitemap index with segmented sitemap files.
Does robots.txt require a sitemap directive?
Not required, but strongly recommended for reliable discovery and tooling clarity.
Sources
- Google Search Central: Build and submit a sitemap
- Google Search Central: Manage your sitemaps with a sitemap index file
- sitemaps.org XML protocol
Need help choosing sitemap segmentation (marketing pages, blog, docs, app pages)? Send support your URL groups.