What is the longest a page can be in the Google Search cache before it is re-crawled?

What is the longest a page can be in the Google Search cache before it is re-crawled? - Google Search Console is a free application that allows you to identify, troubleshoot, and resolve any issues that Google may encounter as it crawls and attempts to index your website in search results. If you’re not the most technical person in the world, some of the errors you’re likely to encounter there may leave you scratching your head. We wanted to make it a bit easier, so we put together this handy set of tips about google-search, googlebot, google-cache, to guide you along the way. Read the discuss below, we share some tips to fix the issue about What is the longest a page can be in the Google Search cache before it is re-crawled?.Problem :


The more popular a page is, the more often Googlebot crawls it.


For less popular pages, Googlebot can go many months between crawls. If the page were taken down in between crawls, the page could remain in the cache for a long time.


What is the longest a stale page can be in the Google Search cache before it gets re-crawled or dropped? Does Google publish this figure?


This is assuming I don't do anything to speed up the removal, like use the Outdated Content tool.


Solution :

It looks like you answered most of this for yourself. It depends how often the page is crawled, which depends on its popularity (mostly). The next biggest factor that may cause problems with crawlers is page depth / click depth. Make sure the pages you want "fresh" can be navigated to within a few clicks from home.


But I don't think it'll ever be months between crawls, even for pages that are very rarely visited. I checked one of my websites that is about 2 months old (domain was fresh too). It only has 4 pages indexed in Google, doesn't update regularly and gets very little traffic, but Googlebot crawls my submitted sitemap almost every day.


So, I think if your sitemap is setup properly, Google will pull the cached copies very quickly without you doing anything. Depending on what you're trying to do, you could use robots meta tags to ask search engines not to offer cached copies in results (noarchive) or simply use Search Console to remove and recrawl.


Last point: The outdated content tool is no longer recommended for site owners. That tool directs you to use "Removals" in Search Console if you can.


If the issue about google-search, googlebot, google-cache, is resolved, there’s a good chance that your content will get indexed and you’ll start to show up in Google search results. This means a greater chance to drive organic search traffic to your site.

Comments

Popular posts from this blog

Years after news site changed name, Google is appending the old name to search titles and news stories

Is it possible to outrank Google for a search term on their own search engine?

Load Wikipedia sourced biographies via Ajax or render it with the rest of the page as part of the initial request?