Tech giant Google has clarified how its Googlebot ranks pages, saying it will crawl the first 15 MB of a webpage and nothing after this cutoff will be included in the ranking calculation.
“After the first 15 MB of a file, Googlebot stops crawling and only considers the first 15 MB of the file for indexing,” Google said.
“The file size limit applies to uncompressed data,” it added.
Reportedly, this left some in the SEO community wondering whether this means that Googlebot will completely ignore text that appears below images at the cutoff in HTML files.
Google search advocate John Mueller clarified via Twitter, “This is specific to the HTML file as written.”
“The embedded resource/content pulled in with the img tag is not part of the HTML file,” he said.
Important content must now be included near the top of web pages to make sure it is weighted by Googlebot.
This means that the code must be structured in a way that holds SEO-relevant information with the first 15 MB in an HTML or supported text-based file.
This also means that images and videos should be compressed, not directly encoded in HTML, whenever possible.
(Only the title and image of this report may have been reworked by Business Standard staff; the rest of the content is generated automatically from a syndicated feed.)