Noindex
Learn what the noindex directive is, how it prevents pages from appearing in search results, and when to use it for better SEO control.
Noindex is a directive that tells search engines not to include a specific page in their search results index. It can be implemented as a meta robots tag in the HTML head or as an X-Robots-Tag HTTP header. When a search engine encounters a noindex directive, it will crawl the page but will not display it in search results, effectively removing it from organic visibility.
Why It Matters for SEO
Not every page on your website should appear in search results. Internal search results pages, login screens, thank-you pages, tag archives, and thin content pages can dilute your site quality in the eyes of search engines. By applying noindex to these pages, you concentrate your site authority on the pages that matter and prevent index bloat from dragging down your overall SEO performance. It also helps manage crawl budget more effectively since search engines learn which pages to deprioritize over time.
How to Implement Noindex
The most common method is adding a meta robots tag in the HTML head section of the page. Alternatively, you can use the X-Robots-Tag HTTP header, which is useful for non-HTML resources like PDFs or images.
Critically, the page must remain crawlable for the noindex directive to work. If you block a page with robots.txt, search engines cannot see the noindex tag and may still index the page based on external signals like backlinks. Always allow crawling of pages you want to deindex.
You can combine noindex with nofollow to both prevent indexing and stop search engines from following links on the page, though these are independent directives that serve different purposes.
Common Mistakes
- Blocking crawling and expecting deindexing: If robots.txt blocks a page, search engines cannot read the noindex tag. Remove the crawl block or use the HTTP header approach.
- Accidentally noindexing important pages: A stray noindex tag on a critical page can cause immediate traffic loss. Audit your site regularly and monitor the Index Coverage report in Google Search Console.
- Including noindexed pages in the sitemap: Your XML sitemap should only contain indexable pages. Sending mixed signals wastes crawl budget.
- Using noindex when canonical is more appropriate: If you have duplicate versions of a page, a canonical URL tag is often more effective than noindex because it consolidates ranking signals rather than discarding them.
- Not verifying after deployment: Always check that your noindex tags are rendering correctly, especially on sites using JavaScript rendering where meta tags may be added dynamically.
Noindex is a powerful tool for controlling what appears in search results, but it must be used deliberately and monitored continuously.