URLs with Noindex Tag in Your Sitemap
A sitemap is your list of URLs you want Google to index. When a page in that list carries <meta name="robots" content="noindex"> or an X-Robots-Tag: noindex HTTP header, you're explicitly contradicting yourself: "please crawl this, but also, do not index it." Google drops these URLs and treats the sitemap as noisier than it should be.
What is this error?
A noindex-in-sitemap error means a URL appears in your sitemap.xml but its page instructs search engines not to index it. Noindex can come from two places: an HTML meta tag (<meta name="robots" content="noindex">) or an HTTP response header (X-Robots-Tag: noindex). Search Console reports these as "Excluded by 'noindex' tag" in the Pages report.
Why does it happen?
Common causes: WordPress pages accidentally set to "Discourage search engines" in Settings → Reading; staging environments pushed to production with noindex still active; SEO plugins (Yoast, Rank Math) applying noindex to category/tag/author archives; CDN rules or server-level X-Robots-Tag headers that weren't updated; and CMS "draft" or "hidden" pages that still appear in auto-generated sitemaps.
Why does it hurt SEO?
Every noindex URL in your sitemap is a wasted entry. Google crawls it, processes the noindex directive, and discards the URL - but still "paid" crawl budget to do so. If a meaningful share of your sitemap (say 10%+) is noindex, Google's trust in the sitemap as a discovery signal drops, meaning your genuinely indexable URLs get crawled less frequently.
How to detect it
Search Console's Pages → "Excluded by 'noindex' tag" section lists all affected URLs, though it may take weeks to surface new cases. Sitemap Fixer fetches each URL and checks both the <meta name="robots"> tag and the X-Robots-Tag header in the HTTP response, giving you the full picture immediately.
How to fix it
1. Decide intent: should this URL be indexed? If yes, remove the noindex tag. If no, remove from sitemap. 2. For WordPress: check Settings → Reading and uncheck "Discourage search engines." 3. For SEO plugins: audit Yoast/Rank Math taxonomy settings for auto-noindex on tags/categories. 4. For server headers: check .htaccess, nginx.conf, or CDN rules for X-Robots-Tag directives. 5. Regenerate the sitemap from a source of truth that excludes noindex pages by default. 6. Resubmit in Search Console and watch the "Excluded by 'noindex'" count drop over 2-3 weeks.
Real-world example
A B2B SaaS relaunched their site and forgot to toggle off WordPress' "Discourage search engines" setting. Their sitemap listed 320 URLs, every one of which carried noindex. Traffic collapsed 87% in two weeks. After removing the site-wide noindex and resubmitting the sitemap, indexing recovered over 6 weeks.
Common mistakes
- Checking only the HTML meta tag and forgetting X-Robots-Tag HTTP headers
- Blocking the URL in robots.txt - Google can't see the noindex if it can't crawl
- Leaving noindex on staging/preview domains that get copied to production