If your business relies on Google Search for customer acquisition, understanding Googlebot’s file size limits is essential to protecting your organic visibility and revenue. Googlebot can crawl up to 15MB of a file, but for Google Search indexing, only the first 2MB of uncompressed HTML content is processed. For PDFs, the indexing limit extends to 64MB. Any HTML content beyond the 2MB threshold is not considered for ranking, which means important text, internal links, or structured data placed too late in the source code may be ignored.
This clarification in Google’s documentation does not introduce a new restriction, but it highlights a technical constraint that can directly impact search visibility if not properly managed.
It is crucial to understand that Google’s February 2026 documentation update did not introduce new technical restrictions. Instead, it reorganized existing information to provide greater clarity between two distinct sets of rules. The first is a universal default applied across all of Google’s various crawlers, which power services beyond just Search, including News, Shopping, and AI systems like Gemini. The second set is specific to Googlebot, the crawler responsible for indexing content for Google Search results.
For years, the 15MB limit was the figure most commonly cited in SEO circles. However, the updated documentation now explicitly states that for the purpose of web search indexing, Googlebot operates under tighter constraints for certain file types. This distinction resolves previous confusion and provides a more accurate framework for technical SEO audits.
To effectively manage your site’s crawlability, you must differentiate between the general infrastructure limit and the Googlebot-specific limits for Search. The table below outlines the key file size thresholds that every SEO professional and web developer must know.
| Crawler Type | File Type | Size Limit | Purpose |
|---|---|---|---|
| All Google Crawlers (General Infrastructure) | Any file type | 15 MB | This is the default maximum file size that any Google crawler will attempt to fetch. It applies to systems powering Google News, AdSense, and other products. |
| Googlebot (For Google Search) | HTML, CSS, JavaScript, and other supported text-based files | 2 MB | When crawling specifically for Google Search, Googlebot will only process the first 2MB of uncompressed data from these files. Content beyond this point is discarded and not sent for indexing. |
| Googlebot (For Google Search) | PDF files | 64 MB | Googlebot can crawl and index a significantly larger portion of PDF documents, up to the first 64MB of uncompressed content. |
This structure highlights a critical point: while the 15MB figure is a system-wide ceiling, the practical limit for your core HTML pages, the foundation of your SEO is a much stricter 2MB when it comes to being indexed in Google Search.
The 2MB limit for HTML files is a direct consequence of Google’s focus on user experience and efficient resource allocation. An HTML file of this size is extraordinarily large; the vast majority of websites operate well below even 200KB. Such a bloated file typically indicates severe underlying issues, such as excessive inline JavaScript and CSS, unoptimized code from a CMS, or an attempt to load an entire application’s worth of content into a single page.
Googlebot stops fetching additional data for indexing consideration once the 2MB uncompressed limit is reached. This means any content, internal links, schema markup, or calls to action that appear after this point are invisible to Google. From an SEO perspective, this is equivalent to that content not existing at all. Furthermore, a page of this size will almost certainly deliver a poor user experience, with slow load times and potential browser instability, which are themselves negative ranking factors.
Ensuring your site complies with these limits is a straightforward process of measurement and optimization.
The easiest way to check your page’s HTML size is through your browser’s Developer Tools.
If your HTML size is approaching or exceeding 500KB, it is time to take action. Key strategies include:
By following these practices, you not only stay safely within Googlebot’s file size limits but also create a faster, more efficient, and user-friendly website. In the current landscape of search, where technical health is a foundational ranking factor, managing your file sizes is not just an SEO tactic it is a core requirement for online visibility.
No. For Google Search indexing, Googlebot processes only the first 2MB of uncompressed HTML content. Any content beyond that limit is not considered for indexing or ranking.
Google’s crawlers have a 15MB default fetch limit for most file types. However, for Google Search indexing, Googlebot applies stricter limits depending on the file type (such as 2MB for HTML and 64MB for PDFs).
No. The file size limits apply to the uncompressed version of the HTML. Even if your page is compressed using gzip or Brotli, Google evaluates the uncompressed size for indexing.
No. The 2MB limit applies to the main HTML document. External resources such as images, CSS, and JavaScript files are requested separately and are not counted toward the HTML size limit.