By SitemapFixer Team
Updated April 2026

Why Are My Pages Not Indexed by Google?

Analyze your sitemap for free Try Sitemap Fixer

Discovering that pages you care about are not in Google's index is one of the most frustrating technical SEO problems. The cause is almost always one of a handful of issues — and the URL Inspection tool in Google Search Console is your primary diagnostic starting point.

Step 1: Use URL Inspection in Google Search Console

Before guessing at the cause, use the URL Inspection tool in Google Search Console. Enter the URL of the page you believe is not indexed and press Enter. This tool gives you the most direct, authoritative view of how Google sees that specific page.

The result will show one of three main statuses:

  • URL is on Google — the page is indexed. If you thought it was not indexed, check whether it ranks for its target keywords using site: search or a rank tracker.
  • URL is not on Google — the page is not indexed. The tool shows the specific reason — blocked by robots.txt, noindex detected, canonical pointing elsewhere, page not found, or server error.
  • URL is on Google, but has issues — the page is indexed but has problems that may limit its performance, such as a mobile usability issue or a soft 404 signal.

The reason shown in URL Inspection is not always the root cause — it is the most visible symptom. Work through the reasons below to understand what is actually causing the issue.

Reason 1: Not in Sitemap and No Internal Links

Google discovers pages in two ways: through your XML sitemap and through crawling links. If a page is absent from your sitemap and no other page on your site links to it, Googlebot has no path to find it. These are called orphan pages.

The fix has two parts. First, add the URL to your sitemap — make sure your sitemap is submitted in Google Search Console and is being fetched successfully. Second, add at least one contextual internal link from a relevant, already-indexed page pointing to the new page. A sitemap alone is not enough: orphan pages receive no PageRank through the internal link graph, which limits how much crawl budget Googlebot allocates to them even when they are in the sitemap.

Reason 2: Blocked by robots.txt

If your robots.txt file contains a Disallow rule that matches the page's URL, Googlebot will not crawl it. The URL Inspection tool will show "Blocked by robots.txt" as the status.

To verify: navigate to https://yoursite.com/robots.txt and look for any Disallow lines that could match your page's path. Use the robots.txt tester in Google Search Console to test the exact URL against your current rules. A common mistake is an overly broad Disallow like Disallow: / (which blocks everything) or a wildcard pattern that accidentally matches more paths than intended.

Important: even after you fix the robots.txt, the page will not be indexed until Googlebot recrawls it. Request indexing via URL Inspection to speed this up.

Reason 3: Noindex Tag or X-Robots-Tag

A noindex meta tag in the page's <head> explicitly tells Google not to index the page. This tag might have been added intentionally (for a staging page or draft post) but forgotten when the page went live. It can also be added unintentionally by a CMS setting.

To check: use URL Inspection and look at the "Coverage" section — it will note if a noindex tag was detected. You can also view the page source (Ctrl+U in most browsers) and search for noindex. Look for <meta name="robots" content="noindex"> in the head section.

For non-HTML files (PDFs, images), servers can send an X-Robots-Tag: noindex HTTP header. Check this with a tool like curl -I https://yoursite.com/page and look for the X-Robots-Tag in the response headers. Both the meta tag and the HTTP header have the same effect — they must both be absent for Google to index the page.

Reason 4: Canonical Points to Different URL

This is the most common hidden cause of non-indexing. If a page has a canonical tag pointing to a different URL, Google indexes the canonical target — not the page you submitted. The page you are investigating may be perfectly accessible and have good content, but Google treats it as a duplicate.

To find it: open URL Inspection for the affected page and expand the Indexing section. It shows the canonical URL that Google detected. If this differs from the page's own URL, you have a canonical mismatch. The GSC coverage status for such pages is typically Alternate page with proper canonical tag.

Common causes: an automated canonical generator producing the wrong URL, a CMS setting the canonical to a paginated version, a trailing-slash inconsistency, or HTTP vs HTTPS mismatch. Fix the canonical to be self-referencing, then request indexing.

Reason 5: Non-200 HTTP Response Code

Google only indexes pages that return a 200 HTTP status code. If a page returns 404 (not found), 301 (redirect), 403 (forbidden), or 5xx (server error), it will not be indexed.

Common situations to check:

  • A page listed in the sitemap that was deleted or moved — returns 404 or 301
  • A URL in the sitemap that redirects to another URL — the redirected URL is what Google indexes, not the sitemap URL
  • Intermittent 503 errors during Googlebot's crawl window — causes the URL to be listed as "Server error (5xx)" in GSC

Verify the HTTP status code using URL Inspection (it shows the last HTTP response) or test manually with curl -I https://yoursite.com/page. Fix the underlying issue — restore the page, update the redirect chain, or resolve the server error — then request indexing.

Reason 6: Thin or Duplicate Content

Google applies quality filters to decide which pages are worth indexing. Pages with very little unique content — short pages that repeat what dozens of other indexed pages say — may be crawled but not indexed. GSC shows these as Crawled — currently not indexed.

To check for near-duplicates, copy a unique sentence from the page and run a Google search with it in quotes. If identical content appears on other indexed pages (including your own), that is a strong signal of the duplicate content problem.

The fix depends on the situation: if the page is genuinely thin, expand it with original, useful content. If it is a functional page (like a tag page or filtered view) that cannot be meaningfully expanded, consider using noindex to exclude it from the index and consolidating its authority via a canonical tag to the main content page.

Reason 7: New Site with Low Crawl Budget

New domains with few or no external backlinks receive very limited crawl budget from Google. It can take weeks to months for all pages on a new site to be discovered and indexed, even when a sitemap is properly submitted.

To accelerate indexing on a new site:

  • Submit your sitemap in Google Search Console and verify it is being fetched
  • Build internal links from already-indexed pages to new pages
  • Earn at least a few external backlinks from indexed, authoritative pages — a single link from a high-authority site can trigger Googlebot to visit your site much sooner
  • Ensure your server is fast — Googlebot crawls more when server response times are low

Typical timeline for a brand-new domain: 4-12 weeks for most pages to be indexed after the first crawl, depending on domain authority and content quality.

How to Request Indexing After Fixing

After identifying and fixing the root cause for a specific page, use the URL Inspection tool to request indexing. Open URL Inspection, enter the URL, and click the Request Indexing button. This adds the URL to Google's crawl queue with a priority signal — it does not guarantee immediate indexing, but it is the fastest way to trigger a recrawl.

Important limits: you can request indexing for individual URLs this way, but this does not work at scale for an entire site. For bulk re-indexing after a large structural change (such as fixing a sitewide canonical issue or removing a sitewide noindex), resubmit your sitemap in the Sitemaps report to signal that your content has changed and should be recrawled.

After requesting indexing, monitor the URL Inspection status over the following 24-72 hours. If the status does not update to "URL is on Google" within a week, there may be a remaining issue blocking indexing that was not addressed by the initial fix.

Check your sitemap now
Free analysis in 60 seconds
Analyze My Sitemap
More guides

Related Guides