Google Indexing Issues: The Complete Troubleshooting Guide 2026
Struggling with Google Search Console indexing? Learn the critical differences between 'Discovered' and 'Crawled - currently not indexed' and how to fix them.
Difference Between “Crawled – Currently Not Indexed” and “Discovered – Currently Not Indexed”
If you have spent any significant amount of time inside Google Search Console (GSC), you have likely encountered the Page Indexing report. For many website owners and SEO managers, this report is a source of anxiety, specifically two status messages that seem almost identical but represent vastly different challenges in the search engine's journey: “Discovered – currently not indexed” and “Crawled – currently not indexed.”
At first glance, the difference might seem semantic. In both cases, the page isn't in Google’s index, and it isn't generating traffic. However, from a technical SEO perspective, these two messages are signals from different stages of Google’s processing pipeline. Understanding which one you are facing is the difference between fixing a "crawl budget" problem and fixing a "content quality" problem.
In this guide, we will break down the mechanics of Google’s crawling and indexing engine, explain why these statuses occur, and provide a systematic diagnostic workflow to move your URLs from "not indexed" to "ranking."
The Anatomy of Google’s Indexing Pipeline: A Deep Dive
To truly understand why a page gets stuck, we must peel back the curtain on how Googlebot actually works. It is not a single, linear process, but a complex, multi-stage operation involving various subsystems that work in parallel.
1. The Scheduler (The Brain)
The scheduler is the brain of the operation. It maintains a massive, persistent database of every URL Google has ever found. For every URL, the scheduler tracks:
- Last Crawl Date: When did we last see this page?
- Crawl Frequency: How often does this content change?
- Priority: Based on internal/external links, how important is this page?
- Crawl Rate Limit: How many requests can we send to this domain without slowing down its server?
When a URL is added to the scheduler (via a sitemap or a link) but hasn't been fetched yet, it resides in the "Discovered – currently not indexed" state.
2. The Fetcher (The Hands)
This is the Googlebot itself. It is a distributed system that makes HTTP requests to servers around the world. It is designed to be polite, respecting robots.txt and the crawl rate limit set by the scheduler. It downloads the raw HTML and handles redirects. If the fetcher encounters a 404 or 5xx error, it reports back to the scheduler.
3. The Processor (The Eyes)
Once the HTML is downloaded, Google’s "Caffeine" indexing system takes over. It parses the HTML to find:
- Metadata: Titles, meta descriptions, robots tags.
- Links: Any
<a href>links are extracted and sent back to the Scheduler to be added to the crawl queue. - Structural Data: Schema.org markup.
- Main Content: The "meat" of the page.
4. The Renderer (The Brain’s Imagination)
Modern websites are rarely just static HTML. They use JavaScript to build the DOM. Because rendering is 100x more expensive than fetching HTML, Google uses a "Two-Wave" model.
- Wave 1: Indexing the raw HTML.
- Wave 2: Rendering the full page with JavaScript.
If your content is only visible after Wave 2, and Wave 2 hasn't happened yet, the page might be classified as "Crawled – currently not indexed" because the first wave saw a "blank" page.
5. The Indexer (The Judge)
Finally, the indexer evaluates the rendered content. It performs deduplication (checking if this page is a copy of another), quality analysis (is this thin content?), and relevance scoring. This is the stage where "Crawled – currently not indexed" decisions are finalized.
1. What “Discovered – Currently Not Indexed” Means
When Google labels a URL as “Discovered – currently not indexed,” it is essentially a "Pending" status. Google knows the URL exists, it's on the "to-do list," but it hasn't started the job yet.
The Prioritization Logic
Why does Google wait? It comes down to Return on Investment (ROI). Every crawl consumes electricity and hardware time. Google wants to spend that time on pages that are likely to provide value to users.
A. The Trust Gap (New Domains)
If your site is new, Google doesn't yet trust that your content is stable or valuable. It will "drip-feed" its crawling, discovering your whole sitemap but only crawling 5–10 pages a day. This is normal and usually resolves with time and consistent publishing.
B. The "Orphan" Signal
One of the loudest signals of "low priority" is an orphan page. If your site structure is a "star" (everything linked from the homepage) versus a "chain" (deeply nested pages), the pages at the end of the chain will stay in "Discovered" much longer. If a page has zero internal links pointing to it, Google assumes you are trying to hide it, or it isn't important enough to show to users.
C. Crawl Budget Exhaustion
For enterprise sites with millions of URLs, crawl budget is a real constraint. If you have a "faceted navigation" problem (thousands of variations of the same product page), Googlebot might spend all its time crawling those variations, leaving your actual new products stuck in the "Discovered" queue.
2. What “Crawled – Currently Not Indexed” Means
This is the "Rejection" status. Google visited the page, and for some reason, decided it shouldn't be in the search results. This is often more frustrating because it implies your content isn't "good enough" or is technically confusing.
The "Quality" Threshold
Google doesn't just index everything. It filters for:
- Unique Value: Does this page provide something that no other page in the index does?
- Technical Health: Does the page load correctly? Are there conflicting meta tags?
- Authority: Is the site as a whole trusted to speak on this topic (E-E-A-T)?
A. Thin and Templated Content
If your page has 100 words of unique text but 2,000 words of boilerplate navigation, footer links, and ads, Googlebot sees it as "thin." This is common on eCommerce product pages where the description is just the manufacturer's provided spec sheet.
B. The Soft 404
A soft 404 is a page that returns a 200 OK status code but looks like an error page. Examples include:
- A search results page with "No items found."
- A category page with zero products.
- A blog post that only says "Coming Soon!" Googlebot is trained to recognize these and will move them to the "Crawled – not indexed" bucket to prevent users from clicking on "dead" links in search results.
C. Canonical Mismatches
This is a technical conflict. You might have a canonical tag pointing to Page A, but your internal links all point to Page B. Google gets confused, crawls both, and might decide to index Page A while putting Page B into "Crawled – not indexed"—or vice-versa.
3. The Key Difference: Direct Comparison
| Factor | Discovered – Currently Not Indexed | Crawled – Currently Not Indexed |
|---|---|---|
| Was it crawled? | No | Yes |
| Content evaluated? | No | Yes |
| Type of problem | Prioritization / Crawl Budget | Quality decision / Relevancy |
| Main solution focus | Crawl signals & Internal linking | Content depth & Content quality |
| Server logs signal | No activity for this specific URL | Successful 200 OK hit recorded |
| Primary driver | Google knows it exists but hasn't visited | Google visited and rejected the content |
4. Advanced Diagnostic: The Architecture of Discovery
To fix "Discovered" issues, you need to understand how Google finds content outside of just following links.
The Role of Sitemaps vs. Feeds
Sitemaps are static. Google checks them occasionally. If you want faster discovery, you should also use RSS or Atom feeds. Googlebot subscribes to these and sees new content almost instantly. If your pages are stuck in "Discovered," check if they are in your RSS feed.
IndexNow: The Instant Discovery Protocol
While not yet adopted by Google, IndexNow (used by Bing and Yandex) allows you to "push" URLs to the search engine. Using IndexNow ensures that at least some search engines see your content immediately, which can sometimes create secondary discovery signals for Google.
Server Logs: The Ground Truth
Don't guess if Google has visited your site. Check your server logs. If you see Googlebot hitting your "Crawled" URLs but GSC says they aren't indexed, you know for a fact the problem is the content, not the access. If you don't see any hits for your "Discovered" URLs, the problem is your architecture.
5. Advanced Diagnostic: The Quality Filter
To fix "Crawled" issues, you need to think like a Quality Rater.
The "Near-Duplicate" Analysis
Google uses a process called "Shingling" to compare pages. It breaks your content into small overlapping sequences of words. If two pages share 80% of the same "shingles," they are duplicates. Fix: Use a tool like 42crawl to run a keyword cannibalization audit. If you have multiple pages targeting the exact same intent, you are forcing Google to choose one and reject the others.
The Rendering Trap
If your site uses "Client-Side Rendering" (CSR), Googlebot sees an empty <div> on the first crawl. If its rendering queue is backed up, it might wait days or weeks to see the actual content. During this gap, the status will be "Crawled – currently not indexed." Fix: Switch to Server-Side Rendering (SSR) or Static Site Generation (SSG) so the content is visible in the raw HTML.
6. Content Pruning vs. Content Improvement: A Decision Framework
When you have a large "Crawled – currently not indexed" bucket, you have two choices: Fix the pages or delete them.
When to Prune (Delete/Redirect):
- The page is an old "Tag" or "Category" that no longer has relevant content.
- The page is a duplicate of a much better page.
- The page has zero organic traffic potential (e.g., "Terms of Service" versions from 2018).
- SEO Benefit: Deleting these pages "reclaims" crawl budget and concentrates your site's authority into fewer, better pages.
When to Improve:
- The page is a product or service you still sell.
- The page targets a keyword with significant search volume.
- The page has high-quality backlinks pointing to it. Fix: Add 500+ words of unique content, add images with descriptive ALT text, and embed a video or a data table.
7. The Role of Site Health in Indexing
Google doesn't just evaluate pages in isolation. It evaluates the Site Health as a whole.
PageRank and Link Equity
Pages with more "votes" (links) are crawled more often and indexed faster. If your important pages are stuck in "Discovered," you likely have a "flat" architecture where everything is equally important, meaning nothing is important. Use internal link analysis to create a hierarchy.
Site Speed and Core Web Vitals
If your site is slow, Googlebot will slow down its crawl rate to avoid crashing your server. This causes the "Discovered" bucket to grow. A fast site is a crawlable site. Check out our Core Web Vitals audit guide to learn how to speed up your site for better indexing.
8. Log File Analysis: A Step-by-Step Guide for Indexing Diagnostics
For technical SEOs, log files are the ultimate source of truth. They show exactly what Googlebot is doing, rather than what GSC says it is doing.
Step 1: Export Your Logs
Access your server logs (Apache, Nginx, or via a CDN like Cloudflare). You are looking for requests where the User-Agent contains "Googlebot".
Step 2: Filter by Status Code
Look for the URLs that GSC has flagged as "Crawled – currently not indexed."
- If you see a 200 OK: Google successfully fetched the page. The issue is purely quality-based.
- If you see a 301 or 302: The page is redirecting. Google might be following the redirect and indexing the target, but leaving the original URL in the "Crawled" bucket.
- If you see a 403 or 429: Your server is blocking Googlebot. This is a critical technical error.
Step 3: Analyze Crawl Frequency
If Googlebot is hitting your "Discovered" URLs but only once every three months, you have a "Crawl Demand" issue. You need to increase the perceived value of those pages by linking to them more frequently from high-traffic areas of your site.
9. Managing Indexing for Large-Scale Migrations
When you migrate a site or launch a massive new section (10,000+ URLs), the "Discovered" bucket will inevitably swell. This is the "Migration Limbo."
The Migration Checklist:
- Warm up the Crawl: Don't push all 10k URLs into the sitemap on day one. Drip-feed them to let Googlebot adjust its crawl rate.
- Redirect Map: Ensure your 301 redirects are 1-to-1. Redirect chains (A -> B -> C) are the fastest way to get your new URLs stuck in the "Crawled – not indexed" bucket.
- Search Console Verification: Verify both the old and new domains (or subfolders) to see the flow of URLs between them.
10. The Impact of AI Search and GEO on Indexing Strategy
In the age of Generative Engine Optimization (GEO), being indexed is no longer the final goal—it is the prerequisite for being cited.
How AI Bots Differ
Bots like GPTBot (OpenAI) and PerplexityBot have different crawl patterns than Googlebot. They are often more aggressive in the "Discovery" phase but more selective in the "Index" phase.
Why Indexing Quality Matters for AI
If your page is in the "Crawled – currently not indexed" bucket, it is highly unlikely to be used as a source for an AI Overview or a chatbot answer. AI engines prioritize the same high-quality, authoritative signals that Google's Indexer does. By fixing your "Crawled" issues, you are simultaneously optimizing for the future of AI-driven search.
11. Systematic Solution with 42crawl
Manual audits are impossible for sites with more than 100 pages. 42crawl provides the automation needed to manage indexing at scale.
Orphan Page Discovery
42crawl matches your crawl data against your sitemap to find URLs that are "invisible" to the site's link structure but visible to Google.
Link Graph Visualization
See your site's architecture as a visual map. Identify "bottlenecks" where Googlebot might get stuck or "dead ends" that prevent discovery of deeper content.
Duplicate Content Detection
42crawl analyzes the similarity between your pages, flagging the "near-duplicates" that cause Google to move pages to the "Crawled – currently not indexed" state.
GEO Readiness Analysis
As search moves toward AI-driven answers, being indexed is just the first step. 42crawl helps ensure your content is structured so that AI bots (like Google's own AI Overviews) can easily "cite" your content once it is indexed.
11. Technical Indexability: The Signals That Control Indexing
Beyond GSC statuses, understanding the technical signals that control indexing is essential for long-term SEO health.
robots.txt: The Gateway
The robots.txt file controls crawling, not indexing directly:
- If a URL is disallowed, Googlebot will not fetch the content
- The Trap: If a page is already indexed and you then disallow it in
robots.txt, Google cannot see thenoindextag. It may remain in the index as a "ghost" result - Best Practice: Ensure critical CSS/JS folders are not disallowed, and use
Disallowfor low-value URL parameters to save crawl budget
Meta Robots Noindex: The Directive
When you want a page to exist but stay out of Search Results:
- For internal search or "Thank You" pages: Use
noindex, follow - For staging environments: Use
noindex, nofollow(or password protection) - Critical: Noindex pages must NOT be blocked by robots.txt, or the directive won't be processed
Canonical Tags: The Hint
The canonical tag tells Google which version is the "preferred" one:
- Important: A canonical is a hint, not a directive. Google may ignore it if pages aren't sufficiently similar
- Conflict: If Page A has a canonical to Page B, but internal links point to Page A, Google gets confused
- The Rule: The URL in an hreflang tag must be the canonical version of that page
HTTP Status Codes: The Response
Understanding status codes is critical:
- 404 (Not Found): "This page doesn't exist" - temporary
- 410 (Gone): "This page is permanently deleted" - leads to faster index removal
- Soft 404: Returns 200 OK but looks like an error - common on empty category pages
Use tools like 42crawl to audit your entire site for these technical signals and ensure no accidental blocks are preventing indexing.
12. Conclusion: The Roadmap to 100% Indexation
The difference between "Discovered" and "Crawled" is the difference between an Access Problem and a Value Problem.
- If it's Discovered: Google can't find it easily or doesn't think it's worth the trip. Fix your internal links and site speed.
- If it's Crawled: Google saw it and wasn't impressed. Fix your content quality and technical signals.
By moving away from "SEO guesswork" and using systematic crawling tools like 42crawl, you can turn the mystery of Google Search Console into an actionable plan for growth. Don't let your best content sit in the "Excluded" bucket—give it the technical foundation it needs to shine in the search results.
Indexing Troubleshooting Checklist
- [ ] Step 1: In GSC, click on "Page Indexing" and identify the largest bucket (Discovered or Crawled).
- [ ] Step 2: Use the "URL Inspection" tool on a sample of 5-10 URLs from each bucket.
- [ ] Step 3: For "Discovered" URLs, check the "Crawl Depth" in 42crawl. If > 3, add more internal links.
- [ ] Step 4: For "Crawled" URLs, check the "Google-selected canonical." If it's different, fix your tags.
- [ ] Step 5: Run a Technical SEO Audit to find hidden "noindex" tags or robots.txt blocks.
- [ ] Step 6: Monitor your Crawl Budget. Are you wasting resources on junk URLs?
- [ ] Step 7: Check your server logs. Is Googlebot hitting the URLs you think it is?
- [ ] Step 8: Evaluate your "Value Proposition." Does this page deserve to rank #1? If not, why should it be indexed?
By following this systematic approach, you can bridge the gap between "Known" and "Indexed," ensuring your content gets the visibility it deserves in both traditional and AI search engines.
Frequently Asked Questions
Related Articles
Meet Your New SEO Teammate: The 42crawl AI Consultant
Discover how we built a lightning-fast AI consultant that understands your website's technical health and provides instant, actionable SEO advice.
Keyword Cannibalization: When Your Best Content is Its Own Worst Enemy
Multiple pages targeting the same intent can tank your rankings. Learn how to detect and resolve keyword cannibalization with 42crawl.
Streamlining SEO Implementation with Jules AI & 42crawl
Discover how direct integration with AI coding agents like Google's Jules can bridge the gap between SEO discovery and technical implementation.