Crawled — Currently Not Indexed: A Coverage Status Guide

Google’s Index Coverage report is totally incredible on the grounds that it gives SEOs more clear bits of knowledge into Google’s creeping and indexing choices. Since its turn out, we use it practically day by day at Go Fish Digital to analyze specialized issues at scale for our customers.

Inside the report, there are a wide range of “statuses” that furnish website admins with data about how Google is dealing with their site content. While a considerable lot of the statuses give some setting around Google’s slithering and indexation choices, one stays hazy: “Crawled — at present not indexed”.

Since seeing the “Crawled — presently not indexed” status reported, we’ve gotten notification from a few site proprietors asking about its significance. One of the advantages of working at an organization is having the option to get before a great deal of information, and on the grounds that we’ve seen this message over different records, we’ve started to get on patterns from reported URLs.

Google’s definition

How about we start with the official definition. As per Google’s legitimate documentation, this status signifies: “The page was crawled by Google, yet not indexed. It could possibly be indexed later on; no compelling reason to resubmit this URL for creeping.”

Along these lines, basically what we can be sure of is that:

  1. Google can get to the page
  2. Google set aside some effort to creep the page
  3. In the wake of creeping, Google chose not to remember it for the index

The way to understanding this status is to consider reasons why Google would “deliberately” rule against indexation. We realize that Google isn’t experiencing difficulty finding the page, yet for reasons unknown it feels clients wouldn’t profit by discovering it.

This can be very disappointing, as you probably won’t know why your content isn’t getting indexed. Beneath I’ll detail the absolute most regular reasons our group has seen to clarify why this strange status may be influencing your website.

1.Bogus positives

Need: Low

Our initial step is to consistently play out a couple of spot checks of URLs hailed in the “Crawled — right now not indexed” area for indexation. It’s normal to discover URLs that are getting reported as barred yet end up being in Google’s index all things considered.

For instance, here’s a URL that is getting hailed in the report for our website:

In any case, when utilizing a site search administrator, we can see that the URL is really remembered for Google’s index. You can do this by adding the content “site:” before the URL.

In case you’re seeing URLs reported under this status, I suggest beginning by utilizing the site search administrator to decide if the URL is indexed or not. Some of the time, these end up being bogus positives.

Arrangement: Do nothing! You’re acceptable.

2.RSS channel URLs

Need: Low

This is one of the most widely recognized models that we see. On the off chance that your site uses a RSS channel, you may be discovering URLs showing up in Google’s “Crawled — at present not indexed” report. Ordinarily these URLs will have the “/feed/” string attached as far as possible. They can show up in the report this way:

Google finding these RSS channel URLs connected from the essential page. They’ll frequently be connected to utilizing a “rel=alternate” component. WordPress modules, for example, Yoast can naturally produce these URLs.

Arrangement: Do nothing! You’re acceptable.

Google is likely specifically deciding not to index these URLs, and all things considered. On the off chance that you explore to a RSS channel URL, you’ll see a XML archive like the one beneath:


While this XML archive is helpful for RSS channels, there’s no requirement for Google to remember it for the index. This would give a poor encounter as the content isn’t intended for clients.

3.Paginated URLs

Need: Low

Another amazingly normal explanation behind the “Crawled — as of now not indexed” avoidance is pagination. We will regularly observe a decent number of paginated URLs show up in this report. Here we can see some paginated URLs showing up from an exceptionally enormous online business site:

Arrangement: Do nothing! You’re acceptable.

Google should slither through paginated URLs to get a total creep of the site. This is its pathway to content, for example, more profound classification pages or item portrayal pages. In any case, while Google utilizes the pagination as a pathway to get to the content, it doesn’t really need to index the paginated URLs themselves.

On the off chance that anything, ensure that you don’t successfully affect the creeping of the individual pagination. Guarantee that the entirety of your pagination contains a self-referential standard tag and is liberated from any “nofollow” labels. This pagination goes about as a road for Google to creep other key pages on your site so you’ll unquestionably need Google to keep slithering it.

4.Lapsed items

Need: Medium

At the point when spot-checking singular pages that are recorded in the report, a typical issue we see across customers is URLs that contain text taking note of “lapsed” or “unavailable” items. Particularly on web based business sites, apparently Google verifies the accessibility of a specific item. In the event that it confirms that an item isn’t accessible, it continues to bar that item from the index.

This bodes well from a UX viewpoint as Google might not have any desire to remember content for the index that clients can’t buy.

Nonetheless, if these items are really accessible on your site, this could bring about a great deal of botched SEO chance. By barring the pages from the index, your content isn’t allowed to rank by any stretch of the imagination.

Also, Google doesn’t simply check the obvious content on the page. There have been cases where we’ve discovered no sign inside the obvious content that the item isn’t accessible. In any case, while checking the organized information, we can see that the “accessibility” property is set to “OutOfStock”.

Apparently Google is taking pieces of information from both the obvious content and organized information about a specific item’s accessibility. Hence, it’s significant that you check both the content and mapping.

Arrangement: Check your stock accessibility.

In case you’re discovering items that are really accessible getting recorded in this report, you’ll need to check the entirety of your items that might be mistakenly recorded as inaccessible. Play out a creep of your site and utilize a custom extraction apparatus like Screaming Frog’s to scratch information from your item pages.

For example, on the off chance that you need to see at scale the entirety of your URLs with outline set to “OutOfStock”, you can set the “Regex” to: “accessibility”:”

This: “class=”redactor-autoparser-object”>” ought to consequently scratch the entirety of the URLs with this property:

You can send out this rundown and cross-reference with stock information utilizing Excel or business insight apparatuses. This ought to rapidly permit you to discover inconsistencies between the organized information on your site and items that are really accessible. A similar procedure can be rehashed to discover examples where your obvious content demonstrates that items are lapsed.

5.301 sidetracks

Need: Medium

One fascinating model we’ve seen show up under this status is goal URLs of diverted pages. Frequently, we’ll see that Google is slithering the goal URL yet excluding it in the index. In any case, after taking a gander at the SERP, we find that Google is indexing a diverting URL. Since the diverting URL is the one indexed, the goal URL is tossed into the “Crawled — as of now not indexed” report.

The issue here is that Google may not be perceiving the divert yet. Therefore, it considers the to be URL as a “copy” since it is as yet indexing the diverting URL.

Arrangement: Create an impermanent sitemap.xml.

In the event that this is happening on countless URLs, it merits finding a way to impart more grounded union signs to Google. This issue could show that Google isn’t perceiving your sidetracks in an ideal way, prompting unconsolidated content signs.

One alternative may be setting up a “brief sitemap”. This is a sitemap that you can make to assist the slithering of these diverted URLs. This is a system that John Mueller has recently suggested.

To make one, you should figure out sidetracks that you have made before:

  1. Fare the entirety of the URLs from the “Crawled — as of now not indexed” report.
  2. Match them up in Excel with diverts that have been recently set up.
  3. iscover the entirety of the sidetracks that have a goal URL in the “Crawled — as of now not indexed” container.
  4. Make a static sitemap.xml of these URLs with Screaming Frog.
  5. Transfer the sitemap and screen the “Crawled — at present not indexed” report in Search Console.

The objective here is for Google to slither the URLs in the transitory sitemap.xml more oftentimes than it in any case would have. This will prompt quicker union of these sidetracks.

6.Slight content

Need: Medium

Now and then we see URLs remembered for this report are very slim on content. These pages may have the entirety of the specialized components set up accurately and may even be appropriately inside connected to, nonetheless, when Google runs into these URLs, there is next to no real content on the page. The following is a case of an item classification page where there is next to no special content:

This item posting page was hailed as “Crawled — Currently Not Indexed”. This might be because of exceptionally dainty content on the page.

This page is likely either unreasonably flimsy for Google to believe it’s helpful or there is so minimal content that Google believes it to be a copy of another page. The outcome is Google expelling the content from the index.

Here is another model: Google had the option to creep a tribute part page in a hurry Fish Digital site (appeared previously). While this content is one of a kind to our site, Google most likely doesn’t accept that the single sentence tribute should remain solitary as an indexable page.

By and by, Google has settled on the official choice to prohibit the page from the index because of an absence of value.

Arrangement: Add progressively content or change indexation signals.

By and by, Google has settled on the official choice to bar the page.

In the event that you accept that the page should be remembered for the index, consider including extra content. This will assist Google with considering the to be as giving a superior encounter to clients.

On the off chance that indexation is pointless for the content you’re finding, the greater inquiry becomes whether you should make the extra moves to firmly flag that this content shouldn’t be indexed. The “Crawled — right now not indexed” report is demonstrating that the content is qualified to show up in Google’s index, however Google is choosing not to incorporate it.

There additionally could be other low quality pages to which Google isn’t holding a candle to the current situation this rationale. You can play out a general “site:” search to discover indexed content that meets indistinguishable measures from the models above. In case you’re finding that countless these pages are showing up in the index, you should consider more grounded activities to guarantee these pages are expelled from the index, for example, a “noindex” tag, 404 mistake, or expelling them from your inner connecting structure totally.

7.Copy content

Need: High

While assessing this prohibition over countless customers, this is the most elevated need we’ve seen. On the off chance that Google considers your to be as copy, it might creep the content however choose not to remember it for the index. This is one of the manners in which that Google maintains a strategic distance from SERP duplication. By expelling copy content from the index, Google guarantees that clients have a bigger assortment of exceptional pages to interface with. Once in a while the report will name these URLs with a “Copy” status (“Duplicate, Google picked unexpected accepted in comparison to client”). Be that as it may, this isn’t generally the situation.

This is a high need issue, particularly on a ton of web based business sites. Key pages, for example, item depiction pages frequently incorporate the equivalent or comparable item portrayals the same number of different outcomes over the Web. On the off chance that Google perceives these as excessively like different pages inside or remotely, it may prohibit them from the index all together.

Arrangement: Add special components to the copy content.

On the off chance that you imagine that this circumstance applies to your site, here’s the manner by which you test for it:

  1. Take a scrap of the expected copy text and glue it into Google.
  2. In the SERP URL, add the accompanying string as far as possible: “&num=100”. This will show you the best 100 outcomes.
  3. Utilize your program’s “Discover” capacity to check whether your outcome shows up in the main 100 outcomes. In the event that it doesn’t, your outcome may be getting sifted through of the index.
  4. Return to the SERP URL and affix the accompanying string as far as possible: “&filter=0”. This should give you Google’s unfiltered result (much appreciated, Patrick Stox, for the tip).
  5. Utilize the “Discover” capacity to look for your URL. In the event that you see your page currently showing up, this is a decent sign that your content is getting sifted through of the index.
  6. Rehash this procedure for a couple of URLs with possible copy or fundamentally the same as content you’re finding in the “Crawled — at present not indexed” report.

In case you’re reliably observing your URLs getting sifted through of the index, you’ll have to find a way to make your content progressively one of a kind.

While there is nobody size-fits-all standard for accomplishing this, here are a few alternatives:

Revise the content to be increasingly one of a kind on high-need pages.

Utilize dynamic properties to naturally infuse exceptional content onto the page.

Expel a lot of superfluous standard content. Pages with more templated text than one of a kind book may be getting perused as copy.

In the event that your site is reliant on client produced content, illuminate benefactors that all given content ought to be extraordinary. This may help forestall occurrences where donors utilize a similar content over numerous pages or spaces.

8.Private-confronting content

Need: High

There are a few occasions where Google’s crawlers access content that they shouldn’t approach. In the event that Google is discovering dev situations, it could remember those URLs for this report. We’ve even observed instances of Google slithering a specific customer’s subdomain that is set up for JIRA tickets. This caused a hazardous slither of the site, which concentrated on URLs that shouldn’t ever be considered for indexation.

The issue here is that Google’s creep of the site isn’t engaged, and it’s investing energy slithering (and possibly indexing) URLs that aren’t intended for searchers. This can have gigantic implications for a site’s slither spending plan.

Arrangement: Adjust your creeping and indexing activities.

This arrangement will be altogether reliant on the circumstance and what Google can get to. Regularly, the main thing you need to do is decide how Google can find these private-confronting URLs, particularly if it’s by means of your inward connecting structure.

Start a slither from the landing page of your essential subdomain and check whether any unwanted subdomains can be gotten to by Screaming Frog through a standard creep. Provided that this is true, it’s protected to state that Googlebot may be finding those identical pathways. You’ll need to evacuate any inner connects to this content to cut Google’s entrance.

The subsequent stage is to check the indexation status of the URLs that ought to be prohibited. Is Google adequately keeping every one of them out of the index, or were some trapped in the index? On the off chance that Google isn’t indexing a lot of this content, you should seriously mull over altering your robots.txt record to square slithering right away. If not, “noindex” labels, canonicals, and secret phrase ensured pages are on the whole on the table.

Contextual analysis: copy client created content

For a genuine model, this is an occasion where we analyzed the issue on a customer site. This customer is like an internet business site as a great deal of their content is comprised of item depiction pages. Nonetheless, these item depiction pages are all client produced content.

Basically, outsiders are permitted to make postings on this site. Be that as it may, the outsiders were regularly adding exceptionally short depictions to their pages, bringing about slender content. The issue happening much of the time was that these client created item portrayal pages were getting trapped in the “Crawled — at present not indexed” report. This brought about botched SEO chance as pages that were equipped for producing natural traffic were totally prohibited from the index.

While experiencing the procedure above, we found that the customer’s item portrayal pages were very slim as far as remarkable content. The pages that were getting prohibited just seemed to have a section or less of novel content. What’s more, the greater part of on-page content was templated text that existed over these page types. Since there was next to no one of a kind content on the page, the templated content may have made Google see these pages as copies. The outcome was that Google rejected these pages from the index, refering to the “Crawled — at present not indexed” status.

To tackle for these issues, we worked with the customer to figure out which of the templated content didn’t have to exist on every item depiction page. We had the option to expel the superfluous templated content from a large number of URLs. This brought about a noteworthy abatement in “Crawled — as of now not indexed” pages as Google considered each to be as progressively exceptional.


Ideally, this helps search advertisers better comprehend the baffling “Crawled — right now not indexed” status in the Index Coverage report. Obviously, there are likely numerous different reasons that Google would decide to sort URLs like this, however these are the most widely recognized examples we’ve seen with our customers to date.

In general, the Index Coverage report is one of the most integral assets in Search Console. I would profoundly urge search advertisers to get acquainted with the information and reports as we routinely find problematic slithering and indexing conduct, particularly on bigger sites. On the off chance that you’ve seen different instances of URLs in the “Crawled — right now not indexed” report, let me know in the remarks!

You have successfully subscribed to the newsletter

There was an error while trying to send your request. Please try again.

CaseyBiggs will use the information you provide on this form to be in touch with you and to provide updates and marketing.