Reduce the Google crawl rate
Google's crawler infrastructure has sophisticated algorithms to determine the optimal crawl rate for a site. Our goal is to crawl as many pages from your site as we can on each visit without overwhelming your server. In some cases, Google's crawling of your site might be causing a critical load on your infrastructure, or cause unwanted costs during an outage. To alleviate this, you may decide to reduce the number of requests made by Google's crawlers.
Understand the cause of the sharp increase in crawling
Sharp increase in crawling may be caused by inefficiencies in your site's structure or issues with your site otherwise. Based on the reports we've received in the past, the most common causes are:
- Inefficient configuration of URLs on the site, which is typically caused by a specific
    functionality of the site: - Faceted navigation or other sorting and filtering functionality of the site
- A calendar with a lot of URLs for specific dates
 
- A Dynamic Search Ad target
We strongly recommend that you check with your hosting company and look at recent access logs of your server to understand the source of the traffic, and see if it fits in the aformentioned common causes of the sharp increase in crawling. Then, check our guides about managing crawling of faceted navigation URLs and optimizing crawling efficiency .
Urgently reduce crawler traffic (for emergencies)
If you need to urgently reduce the crawl rate for short period of time (for example, a couple
  of hours, or 1-2 days), then return 500 
, 503 
, or 429 
HTTP
  response status code instead of 200 
to the crawl requests. Google's crawling
  infrastructure reduces your site's crawling rate when it encounters a significant number of URLs
  with 500 
, 503 
, or 429 
HTTP response status codes (for
  example, if you disabled your website 
).
  The reduced crawl rate affects the whole hostname of your site (for example, subdomain.example.com 
), both the crawling of the URLs that return errors, as well as
  the URLs that return content. Once the number of these errors is reduced, the crawl rate will
  automatically start increasing again.
Exceptional requests to reduce crawl rate
If serving errors to Google's crawlers is not feasible on your infrastructure, file a special request to report a problem with unusually high crawl rate, mentioning the optimal rate for your site in your request. You cannot request an increase in crawl rate, and it may take several days for the request to be evaluated and fulfilled.

