Google Just Slashed Crawl Limits to 2MB from 15MB
Google has changed the way it talks about reading your website but the reality is less scary than the headline suggests.
Before February 5, 2026, the documentation said Googlebot would grab the first 15MB of a file but now they have clarified that for HTML files specifically, the limit for indexing is actually 2MB.
This means if your code is bloated and your important content sits past that 2MB mark, Google simply won’t see it for ranking purposes. It stops reading. That is the long and short of it. You probably don’t need to panic unless you are running a site with catastrophic code bloat.
The headline is misleading
I have been doing SEO for a long time. I’m 37 now and I have seen panic sweep through this industry more times than I care to count.
The recent news about the Google crawl limit dropping from 15MB to 2MB feels like one of those moments where everyone loses their heads a bit. It’s understandable.
When you hear “slashed” you think of a budget cut or a disaster. But here is what actually happened. Google reorganized their documentation. They didn’t wake up one morning and decide to destroy the internet’s ability to rank long pages. They clarified a distinction that has likely existed for a while.
The 15MB limit still exists. It applies to the general crawling infrastructure. That covers everything. Search & News & Shopping.
But the 2MB limit is specific to the Google Search crawler processing HTML. They are just being more precise about which limit applies to which specific process. It is a documentation cleanup rather than a sudden destruction of capacity.
I suspect this distinction has been there in the background for years. We just didn’t have it written down in black and white. Now we do.
How big is 2MB really
Let’s look at the math because numbers usually calm people down. Or scare them. Hopefully the former in this case. Two megabytes of text is massive. Absolutely massive.
If you take a standard novel, say something around 400 pages, that is roughly what 2MB of text looks like. It is about 2 million characters.
Ask yourself a question. Does your homepage contain the entire text of a 400-page novel? I certainly hope not. If it does, you have bigger problems than a Google crawl limit.
Most websites are nowhere near this. According to data from the Web Almanac, the median mobile homepage is something like 22KB. That is 0.022MB. To hit the new 2MB ceiling you would need to be roughly 90 times larger than the average page. It is hard to do that by accident. You really have to try.
So for 99% of the clients I work with at Breakline, this update changes absolutely nothing. Their pages are light. They are fast. Google reads them just fine.
Fetching vs Indexing
This is the technical bit where people get confused. There is a difference between Google fetching a file and Google indexing a file. John Mueller from Google made this point recently and it is worth repeating.
Fetching is the act of downloading the file from your server. The 15MB limit applies there. Googlebot can download up to 15MB. That part hasn’t changed.
But indexing is the part where Google tries to understand what is on the page to rank it. That is where the 2MB limit kicks in.
If your file is 5MB, Google might fetch the whole thing. But it will only “read” the first 2MB for the purpose of ranking you in search results. The rest is essentially invisible text. It exists, but Google ignores it.
It seems harsh. But it makes sense from an engineering standpoint. Why would a search engine need to process 10MB of code to understand if a page is about plumbing services in London? It wouldn’t.
If you haven’t made your point in the first 2 million characters, you probably aren’t going to make it in the next 2 million.
Who actually needs to worry
I said most people are safe. That is true. But “most” isn’t “all”. There are specific scenarios where this 2026 update is going to cause headaches. I have seen some messy sites in my 15 years and I know exactly the type of setups that are at risk.
Legacy CMS platforms are the usual suspects. I am talking about old, custom-built systems that have been patched together by five different developers over a decade.
They often dump huge amounts of inline code right into the HTML. Instead of having a nice clean external CSS file, they write thousands of lines of style information directly into the page source. That counts towards your 2MB. Every character of code counts.
Another culprit is data URLs. This is when you take an image and encode it as a string of text (base64) directly in the HTML instead of linking to an image file. It looks like a massive block of gibberish code.
If you have a few high-resolution images encoded like this, you can hit 2MB instantly. I honestly don’t know why people still do this for large images but they do.
If you are using a heavy WordPress theme that is poorly coded, you might also be creeping up. Some visual builders add an incredible amount of bloat. It is just div inside div inside div for miles.
How to check your size
You might be sitting there thinking “I have no idea how big my HTML is”. That is fair. It is not something we usually check daily. But you should check it now just to be sure.
You can use the Developer Tools in your browser. I use Chrome mostly. Right-click on your page and hit Inspect. Go to the Network tab. Refresh the page. You will see a list of all the files loading. Filter by “Doc”. That shows you the main HTML document.
Here is the trick though. You need to look at the uncompressed size. The network tab often shows you the transfer size, which is smaller because of compression like GZIP or Brotli. Google looks at the uncompressed version. The raw bytes. You need to make sure that number is under 2MB.
You can also use tools like Screaming Frog. It has a column for size. It is pretty reliable. If you see anything approaching 1MB, I would start asking questions.
If you are over 1.5MB, you are in the danger zone. You need to accomodate for some margin of error. Yes I know it is annoying to add another step to your audits. But better safe than sorry.
The JavaScript problem
There is another layer to this onion. JavaScript.
Modern web development loves JavaScript. We build everything with it. But search engines have a complicated relationship with it. When Google crawls a page, it often grabs the initial HTML first. Then, if it has resources, it renders the JavaScript to see the final version of the page.
The 2MB limit applies to the HTML file. If your important content is injected via JavaScript, and your initial HTML is bloated with script tags, you might be pushing important stuff out of view.
Also, consider the AI aspect. We are in 2026 now. AI search engines and LLM crawlers are everywhere. Many of them do not render JavaScript as well as Google does. They often just grab the raw HTML.
If your content is buried or if your file is too big, they might just skip you. They want efficiency. They don’t want to parse 5MB of spaghetti code to find a paragraph of text. I think this is where the industry is heading. Leaner is better. Clean code wins.
Simple ways to reduce bloat
If you find yourself over the limit, don’t panic. Fixing this is usually straightforward. It just takes some work.
First, move your scripts and styles to external files. There is almost no reason to have thousands of lines of CSS in your HTML head. Put it in a .css file. Google fetches that separately and it has its own cache and limits. It cleans up your HTML immediately.
Second, stop using base64 for images. Just use standard image tags. It is faster for users anyway.
Third, look at your tracking codes. I have audited sites that have fifty different tracking scripts in the header. Facebook pixels & LinkedIn insights & Hotjar & five other tools the marketing team forgot about three years ago.
Remove the ones you don’t use. It speeds up your site and lowers your file size. It is basically spring cleaning. We all hoard junk in our code. This update is just a reminder to take the trash out.
Final Thoughts
I want to be clear about one thing. This is not a crisis. It is a calibration. Google is telling us how their system works so we can build better websites. That is actually helpful.
When I started in this business, we were guessing half the time. Now they give us specific numbers. 2MB. That is a target. It is a generous target. If you can’t say what you need to say in 2MB, you are probably talking too much.
Be skeptical of the fear-mongering. The agencies screaming that “SEO is dead” or “Google killed crawling” are just trying to sell you an audit.
You probably do need an audit, but not because the sky is falling. You need it because good hygiene matters.
Keep your content high up in the HTML structure. Don’t bloat your site with unnecessary junk. It is the same advice I would have given you five years ago. It is just more relevant today.
