By SitemapFixer Team
Updated April 2026

Sitemap Errors in Google Search Console: Complete Fix Guide

Diagnose your sitemap freeAnalyze Free

Google Search Console surfaces sitemap errors that range from simple fetch failures to subtle indexing signals. Understanding what each error means — and which ones actually block indexing — lets you prioritize fixes and recover rankings faster.

Could Not Fetch Sitemap

Googlebot could not download your sitemap file. Check your sitemap URL returns a 200 status, is not blocked by robots.txt, requires no authentication, and loads in under 10 seconds. You can also catch these issues before Google does using a sitemap checker that validates every URL independently.

Sitemap is an HTML Page

Your sitemap URL returns an HTML page instead of XML. Common cause: your CMS returns a 404 error page at /sitemap.xml. Ensure your sitemap generation plugin is active and generating valid XML.

Invalid XML

Your sitemap contains malformed XML. Common causes: unencoded special characters, missing closing tags, incorrect encoding. Validate with an XML validator before submitting.

URLs in Sitemap Cannot Be Crawled

Google can reach your sitemap but cannot crawl the listed URLs. Check robots.txt rules blocking the URLs, server errors when fetching individual pages, and pages behind authentication.

Submitted URL Blocked by robots.txt

A URL in your sitemap is blocked by your robots.txt. Either remove the URL from your sitemap or fix your robots.txt to allow Googlebot access.

Sitemap Indexed Count vs Submitted Count Gap

A large gap between URLs submitted and URLs indexed is not an error but a warning. It means Google is choosing not to index many of your submitted pages. Focus on content quality improvements for the non-indexed pages.

Submitted URL Not Selected as Canonical

This status means Google crawled your URL but decided a different URL is the canonical version. It won't index the URL you submitted. Common triggers are duplicate content served at multiple URLs, inconsistent internal linking pointing to a different variant, or a self-referencing canonical tag that doesn't match the URL exactly. Fix by ensuring the page's canonical tag matches the URL in your sitemap, and that all internal links point to the same canonical form.

Submitted URL Has Noindex

A page in your sitemap carries a noindex directive — either in a meta robots tag or an X-Robots-Tag HTTP header. Google will not index it and will report it as an error. The fix is to either remove the noindex directive if you want the page indexed, or remove the URL from your sitemap if the noindex is intentional. Leaving noindex pages in a sitemap creates conflicting signals and wastes crawl budget.

Soft 404 Detection

A soft 404 occurs when a page returns HTTP 200 but Google determines the content signals a missing or empty page — for example, a "product not found" page or a thin auto-generated page with almost no body text. Google reports these under Coverage in Search Console. Fix by either adding real content to the page or returning a proper 404 or 410 status code and removing the URL from your sitemap.

Too Many URLs Warning

A single sitemap file may contain at most 50,000 URLs and must be under 50MB uncompressed. If you exceed these limits, Google may truncate processing and miss URLs at the end of the file. Split large sitemaps into smaller child files and reference them through a sitemap index file. Prioritize your most important URLs in each child sitemap rather than ordering by publication date alone.

How to Prioritize Which Errors to Fix First

Not all sitemap errors have equal impact. Fix in this order: first resolve any errors that block Google from fetching the sitemap at all (HTTP errors, robots.txt blocks); then fix noindex conflicts which are direct indexing blockers; then address canonical mismatches for your highest-traffic pages; finally work through soft 404s and content quality issues. The Coverage report in GSC lets you filter by error type and sort by page count to find the highest-leverage fixes.

Using URL Inspection to Diagnose Individual URL Issues

For any URL showing an unexpected status in the Coverage report, run it through the URL Inspection tool in GSC. It shows the last crawl date, the crawled content, the detected canonical, any noindex signals, and whether the page is eligible for indexing. You can also trigger a fresh crawl using "Request Indexing" after making fixes. This is the fastest way to confirm a fix worked without waiting for the next full crawl cycle.

Related Guides

Fix your sitemap now
Free analysis in 60 seconds
Analyze My Sitemap
Related guides