Google Not Indexing Pages — Why It Happens and How to Fix It

When Google is not indexing your pages, those pages cannot appear in search results. It is a fundamental problem that blocks all other SEO efforts — no amount of content quality, link building, or on-page optimisation matters if the page is not in Google’s index. The good news is that indexing problems are almost always diagnosable and fixable once you understand the cause.

This guide covers the most common reasons Google does not index pages and what to do about each one. For a broader look at how crawling and indexing work, see the guide on crawlability and indexing.

How to check if your pages are indexed

Before diagnosing the cause, confirm which pages are and are not indexed. There are two main ways to do this:

The most reliable method is Google Search Console’s Coverage or Indexing report, which categorises all known URLs into indexed, excluded, and errored states — with a reason for each exclusion. This is the primary tool for understanding indexation at scale.

For individual pages, the URL Inspection tool in Search Console shows the current indexed status, the last time the page was crawled, and what Googlebot saw when it visited. It also lets you request a fresh crawl after making changes.

The page has a noindex tag

A noindex directive tells Google explicitly not to include the page in its index. It can be set as a meta robots tag in the page’s HTML (meta name=’robots’ content=’noindex’) or as an X-Robots-Tag in the HTTP response header.

Noindex tags are sometimes added accidentally — by a WordPress plugin, a CMS setting applied during development, or a migration that did not fully reverse a staging configuration. Check the page source or use a browser extension to inspect the robots meta tag before investigating further causes.

The page is blocked by robots.txt

Robots.txt is a file that instructs crawlers which parts of a site they are allowed to access. If a page or directory is disallowed in robots.txt, Googlebot will not crawl it — and an uncrawled page cannot be indexed.

This is distinct from noindex: a robots.txt block prevents crawling entirely, while noindex prevents indexing of a crawled page. Importantly, a page blocked by robots.txt that also has a noindex tag will not have the noindex respected — because the noindex tag can only be read by crawling the page.

Check your robots.txt file directly at yourdomain.com/robots.txt or use Google Search Console’s robots.txt tester to confirm whether a specific URL is being blocked.

The page is set to canonical to a different URL

A canonical tag tells Google which version of a page is the definitive one. If a canonical tag points to a different URL, Google treats the current page as a duplicate and indexes the canonical target instead.

Canonical tag errors are extremely common. They include canonicals pointing to the wrong URL due to a template error, self-referencing canonicals that are missing or incorrect, and canonicals set by a plugin that override intentional configuration. Check the canonical tag in the page source or via the URL Inspection tool.

The page has thin or low-quality content

Google actively filters out pages it considers to provide insufficient value to searchers. This includes pages with very little content, pages that are near-identical to other pages on the same site, and pages that appear to be generated for SEO purposes rather than for users.

If a page has been crawled but Google has chosen not to index it, and there are no technical barriers, content quality is a likely factor. The fix is to substantially improve the page’s depth and usefulness — adding meaningful content that addresses the searcher’s intent better than existing indexed pages do.

The page has no internal links pointing to it

Google discovers pages by following links. A page with no internal links pointing to it — an orphan page — has no path for Googlebot to find it through normal crawling. While sitemaps can surface these pages, a page with no internal links is also sending a signal that it is not important enough to link to — which can influence indexation decisions.

The fix is straightforward: add internal links from relevant, already-indexed pages. The anchor text used in those links should reflect what the target page is about.

The page is too new

New pages — especially on new or low-authority sites — are not always indexed immediately. Google crawls and indexes the web at its own pace, prioritising pages on sites it considers authoritative and up-to-date. A brand new page on an established site may be indexed within hours. The same page on a new site may take days or weeks.

You can request indexing via the URL Inspection tool in Search Console, which puts the page in a priority crawl queue. This does not guarantee immediate indexing but is a useful first step.

The server returned an error

If the server returns a 5xx error when Googlebot tries to crawl a page, the page cannot be indexed. Persistent server errors on key pages should be treated as a high-priority technical issue. Check your server logs and hosting status if Search Console is reporting server errors on important URLs.

What to do if pages remain unindexed after fixes

After addressing the likely cause, use the URL Inspection tool to request indexing and monitor Search Console’s Coverage report over the following days and weeks. Some changes — particularly improvements to content quality — take longer for Google to re-evaluate than simple configuration fixes.

If you have gone through these steps and pages are still not being indexed, it may be a more complex issue — a site-wide crawl budget problem, a rendering issue affecting how Google processes the page, or a pattern of quality signals across the site that is suppressing indexation more broadly. In these situations, a structured technical SEO audit is the most efficient way to identify the root cause.

For ongoing technical support, find out more about my technical SEO services.

Scroll to Top