Back to blog

Technical

robots.txt and noindex Basics

robots.txt and noindex serve different jobs and must not be mixed.

January 31, 20265 min readLuca Bell
robots-txtnoindexcrawl-control

Use robots for crawl paths

Block private or duplicate paths, but keep valuable URLs crawlable.

Why this matters

Technical quality protects every SEO and conversion initiative. If rendering, caching, and crawl directives are inconsistent, content quality alone cannot unlock growth.

Implementation checklist

  • Document route behavior before changing render or cache settings.
  • Ship changes behind measurable checks (logs, alerts, and audits).
  • Validate canonical, robots, and status-code behavior in staging.
  • Create rollback steps for cache and routing changes.

Common mistakes to avoid

  • Changing multiple infrastructure variables in the same release.
  • Relying on lab metrics only and ignoring field data.
  • Treating cache invalidation as a manual afterthought.

Use noindex for visibility control

Apply noindex to pages you want crawled but excluded from results.

Why this matters

Technical quality protects every SEO and conversion initiative. If rendering, caching, and crawl directives are inconsistent, content quality alone cannot unlock growth.

Implementation checklist

  • Document route behavior before changing render or cache settings.
  • Ship changes behind measurable checks (logs, alerts, and audits).
  • Validate canonical, robots, and status-code behavior in staging.
  • Create rollback steps for cache and routing changes.

Common mistakes to avoid

  • Changing multiple infrastructure variables in the same release.
  • Relying on lab metrics only and ignoring field data.
  • Treating cache invalidation as a manual afterthought.

Test before deployment

Validate critical URLs in Search Console before publishing rule changes.

Why this matters

Technical quality protects every SEO and conversion initiative. If rendering, caching, and crawl directives are inconsistent, content quality alone cannot unlock growth.

Implementation checklist

  • Document route behavior before changing render or cache settings.
  • Ship changes behind measurable checks (logs, alerts, and audits).
  • Validate canonical, robots, and status-code behavior in staging.
  • Create rollback steps for cache and routing changes.

Common mistakes to avoid

  • Changing multiple infrastructure variables in the same release.
  • Relying on lab metrics only and ignoring field data.
  • Treating cache invalidation as a manual afterthought.

Final takeaway

Small directive mistakes can hide entire site sections from search.

Metrics to monitor

  • Crawl success rate
  • Cache hit ratio by route
  • LCP/INP field data
  • Indexed vs submitted URL gap

Related articles

Updated February 7, 2026 https://www.seorender.io/en/blog/robots-txt-and-noindex-basics