Google Search Console’s indexing and coverage systems form the diagnostic core of technical SEO. This pillar explains how Google discovers URLs, evaluates them for indexing, flags issues, and reports crawl behavior. A clean indexing foundation ensures that content is eligible to rank, that Googlebot can access resources, and that search performance data reflects the true state of your site.
How Google Discovers & Evaluates URLs
Google uses multiple mechanisms to find and evaluate pages:
- Sitemaps submitted in GSC
- Internal links
- External backlinks
- URL Inspection requests
- Canonical signals
- JavaScript-rendered links
- Server logs and historical crawl data
Once discovered, URLs enter Google’s crawl → render → index → serve pipeline. GSC’s Coverage and Page Indexing reports reveal where in this pipeline issues occur.
Page Indexing Report
The Page Indexing report shows which URLs Google has indexed and why others were excluded. It categorizes URLs into:
- Indexed — eligible to appear in search
- Not indexed — excluded for specific reasons
- Blocked — intentionally or unintentionally restricted
- Crawled but not indexed — discovered but not selected for indexing
Each status reflects a different technical or quality signal.
Common Indexing Statuses & What They Mean
GSC provides detailed explanations for each exclusion type. The most important include:
- Discovered – currently not indexed — Google knows the URL exists but hasn’t crawled it yet, often due to crawl budget or low priority.
- Crawled – currently not indexed — Google crawled the page but chose not to index it, usually due to quality or duplication.
- Duplicate without user‑selected canonical — Google found similar content and chose a different canonical.
- Alternate page with proper canonical — Google indexed the canonical version instead.
- Blocked by robots.txt — crawling prevented by robots rules.
- Soft 404 — page returns content that looks like an error.
- Server error (5xx) — hosting or backend issues preventing crawling.
Understanding these statuses is essential for diagnosing indexing gaps.
Crawl Stats Report
The Crawl Stats report provides visibility into Googlebot’s activity:
- Total crawl requests
- Crawl response types
- File types crawled
- Host status
- Crawl purpose (refresh vs. discovery)
Spikes or drops in crawl activity often indicate technical issues, migrations, or major content changes.
URL Inspection Tool
URL Inspection is the most precise diagnostic tool in GSC. It shows:
- Whether a URL is indexed
- Canonical selected by Google
- Last crawl date
- Crawlability and indexability signals
- Rendered HTML
- Page resources and JavaScript behavior
It also allows Request Indexing, which can accelerate discovery for new or updated content.
Sitemaps & Indexing Efficiency
Sitemaps help Google prioritize crawling. Best practices include:
- Submitting a sitemap index
- Using separate sitemaps for large sections
- Ensuring only canonical URLs are included
- Removing 404s, redirects, and parameter URLs
- Regenerating sitemaps automatically on updates
Clean sitemaps improve crawl efficiency and reduce indexing delays.
Diagnosing Indexing Problems
Most indexing issues fall into a few categories:
- Quality issues — thin content, duplication, low value
- Technical issues — blocked resources, server errors, JS rendering failures
- Canonical conflicts — inconsistent signals across pages
- Crawl budget limitations — large sites with many low‑value URLs
- Incorrect robots or meta directives — accidental noindex or disallow rules
GSC’s reports help isolate the root cause quickly.
Why This Pillar Matters
Indexing determines:
- Whether your content can rank
- How much of your site Google sees
- How efficiently Googlebot crawls your pages
- How quickly new content appears in search
- How accurately performance data reflects reality
A strong indexing and coverage foundation ensures that all future SEO work—content, links, performance, schema—can actually produce results.