Written on 17/3/2026
Updated on 19/3/2026

Noindex: when and how to exclude a page from Google results

Definition

The noindex directive is an instruction you give Google asking it not to include a page in its search results. It's not a defensive tool: it's a precision tool for protecting your site's overall quality and concentrating authority on pages that deserve to be seen.

What is noindex?

The noindex directive is a technical SEO instruction telling search engines not to include a page in their index. It's placed in the <meta name="robots" content="noindex"> tag in the page's <head>, or in the HTTP header X-Robots-Tag: noindex. A noindex page can be crawled by Googlebot, but it won't be shown in search results and doesn't contribute to the site's overall quality.

Why and when to use noindex in 2026

Noindex isn't only a security tool: it's a lever for managing the site's overall quality. Google evaluates a site's quality partly through the proportion of low-value pages in its index. A site with many thin, duplicated, or useless pages sees its overall quality penalized. By noindexing these pages, you concentrate the quality signal on your important pages. Pages typically to exclude: e-commerce filter pages, pagination, tag pages, internal search results, thank-you pages, login pages, duplicate content pages.

What we observe at Vydera on misused noindex

The most frequent error in audits: noindex directives accidentally applied to strategic pages. This often happens during migrations, template changes, or SEO plugin updates. The page remains accessible to users, but Google no longer shows it in results. The problem: traffic drops progressively without any apparent error message. Checking robot directives on key pages is an indispensable step in any SEO audit.

How to implement and verify noindex

  • Implement noindex via the meta robots tag in the page head or via your SEO tool (Yoast, RankMath, Webflow SEO settings).
  • Verify your directives via the URL Inspection tool in Google Search Console: it shows whether a page is indexed and why.
  • Don't confuse noindex with Disallow in robots.txt: Disallow blocks crawling but can leave the page indexed. Noindex allows crawling but prevents indexing. To properly exclude, use noindex (not Disallow).

Sources and references

Go further

Robot directives are part of every technical audit. Find our analyses on Vydera Lab or contact us for an index coverage audit of your site.

No, not immediately. Google must first recrawl the page to detect the new noindex directive, then remove it from its index. This process can take days to several weeks depending on your site's crawl frequency. To accelerate the process, you can use the URL Inspection tool in Google Search Console and request a recrawl.

Technically, a noindex page can still be crawled, meaning the links it contains can be followed by Googlebot. But a de-indexed page has reduced authority: it will no longer receive organic traffic, therefore no engagement signals. In practice, noindex pages transmit little value. If you want to keep a page's links without indexing it, it's a situation worth careful consideration.

These are two different directives acting at different stages. Disallow in robots.txt prevents Googlebot from crawling the page: it can't read it. Noindex in the meta tag allows crawling but instructs not to index. Combining both is sometimes used but is counterproductive: if Google can't crawl the page, it can't read the noindex directive and may still index it from external links.

Not by default. The noindex directive is specific to search engines (Google, Bing). AI robots like GPTBot, PerplexityBot, or ClaudeBot don't respect the meta noindex tag: they rely on the robots.txt file. To exclude your content from LLMs' RAG systems, you need to add specific rules per User-Agent in your robots.txt.