URLs with Noindex Tag in Your Sitemap

Updated April 2026·By SitemapFixer Team

A sitemap is your list of URLs you want Google to index. When a page in that list carries <meta name="robots" content="noindex"> or an X-Robots-Tag: noindex HTTP header, you're explicitly contradicting yourself: "please crawl this, but also, do not index it." Google drops these URLs and treats the sitemap as noisier than it should be.

Find noindex URLs in your sitemap
We check both the meta tag and HTTP headers for every URL
Analyze My Sitemap

What is this error?

A noindex-in-sitemap error means a URL appears in your sitemap.xml but its page instructs search engines not to index it. Noindex can come from two places: an HTML meta tag (<meta name="robots" content="noindex">) or an HTTP response header (X-Robots-Tag: noindex). Search Console reports these as "Excluded by 'noindex' tag" in the Pages report.

Why does it happen?

Common causes: WordPress pages accidentally set to "Discourage search engines" in Settings → Reading; staging environments pushed to production with noindex still active; SEO plugins (Yoast, Rank Math) applying noindex to category/tag/author archives; CDN rules or server-level X-Robots-Tag headers that weren't updated; and CMS "draft" or "hidden" pages that still appear in auto-generated sitemaps.

Why does it hurt SEO?

Every noindex URL in your sitemap is a wasted entry. Google crawls it, processes the noindex directive, and discards the URL - but still "paid" crawl budget to do so. If a meaningful share of your sitemap (say 10%+) is noindex, Google's trust in the sitemap as a discovery signal drops, meaning your genuinely indexable URLs get crawled less frequently.

How to detect it

Search Console's Pages → "Excluded by 'noindex' tag" section lists all affected URLs, though it may take weeks to surface new cases. Sitemap Fixer fetches each URL and checks both the <meta name="robots"> tag and the X-Robots-Tag header in the HTTP response, giving you the full picture immediately.

How to fix it

1. Decide intent: should this URL be indexed? If yes, remove the noindex tag. If no, remove from sitemap. 2. For WordPress: check Settings → Reading and uncheck "Discourage search engines." 3. For SEO plugins: audit Yoast/Rank Math taxonomy settings for auto-noindex on tags/categories. 4. For server headers: check .htaccess, nginx.conf, or CDN rules for X-Robots-Tag directives. 5. Regenerate the sitemap from a source of truth that excludes noindex pages by default. 6. Resubmit in Search Console and watch the "Excluded by 'noindex'" count drop over 2-3 weeks.

Real-world example

A B2B SaaS relaunched their site and forgot to toggle off WordPress' "Discourage search engines" setting. Their sitemap listed 320 URLs, every one of which carried noindex. Traffic collapsed 87% in two weeks. After removing the site-wide noindex and resubmitting the sitemap, indexing recovered over 6 weeks.

Common mistakes

Frequently Asked Questions

Why would a URL have noindex but still be in the sitemap?
Usually it's unintentional - staging pages that went live with their noindex tag intact, plugin defaults that add noindex to archive/tag/author pages, or dynamic sitemap generators that don't check the meta robots tag before including a URL.
Will Google penalize my sitemap for including noindex URLs?
Not directly, but it's a quality signal. John Mueller has said sitemaps should list pages you want indexed - noindex URLs waste crawl budget and reduce the signal quality of the entire sitemap.
Is noindex in a meta tag the same as X-Robots-Tag: noindex?
Yes, they have identical effects. The HTTP header version (X-Robots-Tag) is useful for non-HTML files like PDFs, while the meta tag works for HTML pages. Both must be checked when auditing a sitemap.
Fix this in your sitemap now
Enter your domain and get a full sitemap audit in 60 seconds
Analyze My Sitemap Free
Related sitemap errors
All sitemap errors