How Google handles JavaScript throughout the indexing process
A study analyzed Googlebot's interaction with JavaScript-heavy sites, revealing it successfully renders all HTML pages and debunking myths about its JavaScript handling, emphasizing modern rendering capabilities and SEO optimization insights.
Read original articlesearch engines. This research aimed to clarify this misconception by analyzing how Googlebot interacts with JavaScript-heavy sites. The study involved over 100,000 fetches from Googlebot, focusing on the rendering success of various pages on nextjs.org, which employs a mix of rendering techniques. The findings revealed that Googlebot successfully rendered all HTML pages, including those with complex JavaScript interactions, and indexed content loaded asynchronously. The research also debunked several myths about Google's handling of JavaScript, including the belief that Google cannot render JavaScript content and that it treats JavaScript pages differently. The analysis showed that Googlebot's rendering capabilities have evolved significantly, now utilizing an up-to-date version of Chrome, which allows it to process modern JavaScript features effectively. The study also examined the impact of rendering queues and timing on SEO, finding that while some pages experienced delays, many were rendered quickly, challenging the notion of a long rendering queue. Additionally, the research indicated that JavaScript-heavy sites do not inherently suffer from slower page discovery. Overall, the study provides valuable insights into optimizing web applications for search engines, emphasizing the importance of understanding Google's current rendering capabilities and debunking outdated beliefs within the SEO community.
Related
Waves of Writing for Google
The article explores the evolution of writing for Google, highlighting shifts from keyword stuffing to user-focused content and AI's impact on writing jobs. Writers are advised to adapt, focus on personal branding, and embrace technology for relevance.
Google has been lying about their search results [video]
A leak from Google's GitHub shows the search algorithm tracks user clicks and time on pages, raising concerns about search result accuracy, treatment of smaller websites, and SEO strategies.
The Cost of JavaScript
JavaScript significantly affects website performance due to download, execution, and parsing costs. Optimizing with strategies like code-splitting, minification, and caching is crucial for faster loading and interactivity, especially on mobile devices. Various techniques enhance JavaScript delivery and page responsiveness.
Google Now Defaults to Not Indexing Your Content
Google has changed its indexing to prioritize unique, authoritative, and recognizable content. This selective approach may exclude smaller players, making visibility harder. Content creators face challenges adapting to Google's exclusive indexing, affecting search results.
'Google says I'm a dead physicist': is the biggest search engine broken?
Google faces scrutiny over search result accuracy and reliability, with concerns about incorrect information and cluttered interface. Despite dominance in the search market, criticisms persist regarding data privacy and search quality.
- Many commenters share personal experiences with JavaScript and SEO, noting challenges with indexing and crawl budgets, especially at scale.
- There is skepticism about the article's conclusions, with some arguing that while Google can render JS, it may not rank those pages effectively.
- Concerns are raised about the impact of bloated JavaScript on SEO and crawl efficiency, advocating for server-side rendering or prerendered HTML.
- Some commenters express a desire for Google to be more transparent about its SEO policies to reduce confusion and improve practices.
- There is a call for better practices in web development, emphasizing the need for accessibility and efficient content delivery.
Although they will happily crawl and render JS heavy content, I strongly suspect bloat negatively impacts the "crawl budget". Although in 2024 this part of the metric is probably much less than overall request latency. If Googlebot can process several orders of magnitude of sanely built pages with the same memory requirement as a single React page, it isn't unreasonable to assume they would economize.
Another consideration would be that "properly" used, a JS heavy page would most likely be an application of some kind on a single URL, whereas purely informative pages, such as blog articles or tables of data would exist on a larger number of URLs. Of course there are always exceptions.
Overall, bloated pages are a bad practice. If you can produce your content as classic "prerendered" HTML and use JS only for interactive content, both bots and users will appreciate you.
HN has already debated the merits of React and other frameworks. Let's not rehash this classic.
Since our core technology is a React app, I realized that we could just mount the React app directly on any path at the customer's domain. I won't get into the exact implementation, but it worked, and our customers' product pages started being indexed just fine. We even ranked competitively with the upstarts who used server-side rendering. We had a prototype in a few months, and then a few months after that we had the version that scaled to 100s of customers.
We then decided to build a new version of our product on Remix (SSR framework similar to nextjs). It required us to basically start over from scratch since most of our technologies weren't compatible with Remix. 2 years later, we still aren't quite done. When all is said and done, I'm really curious to see how this new product SEOs compared to the existing one.
The same outcome is gonna happen either way, Google will say what their policy is, or people will spend time and bandwidth figuring out their policy. Either way, Google's policy becomes public.
Google could even come out and publish stuff about how to have good SEO, and end all those scammy SEO help sites. Even better, they could actively try to promote good things like less JS when possible and less ads and junk. It would help their brand image and make things better for end users. Win-win.
Also, competition. In a highly competitive seo environment like US real estate, we were constantly competing with 3 or 4 other well-funded and motivated companies. A couple times we tried going dynamic first with a page we lost rankings. Maybe it’s because fcp was later? I don’t know. Because we ripped it all out and did it server side. We did use NextJs when rebuilding trulia but it’s self hosted and only uses ssr.
Please only use JavaScript for dynamic stuff.
Related
Waves of Writing for Google
The article explores the evolution of writing for Google, highlighting shifts from keyword stuffing to user-focused content and AI's impact on writing jobs. Writers are advised to adapt, focus on personal branding, and embrace technology for relevance.
Google has been lying about their search results [video]
A leak from Google's GitHub shows the search algorithm tracks user clicks and time on pages, raising concerns about search result accuracy, treatment of smaller websites, and SEO strategies.
The Cost of JavaScript
JavaScript significantly affects website performance due to download, execution, and parsing costs. Optimizing with strategies like code-splitting, minification, and caching is crucial for faster loading and interactivity, especially on mobile devices. Various techniques enhance JavaScript delivery and page responsiveness.
Google Now Defaults to Not Indexing Your Content
Google has changed its indexing to prioritize unique, authoritative, and recognizable content. This selective approach may exclude smaller players, making visibility harder. Content creators face challenges adapting to Google's exclusive indexing, affecting search results.
'Google says I'm a dead physicist': is the biggest search engine broken?
Google faces scrutiny over search result accuracy and reliability, with concerns about incorrect information and cluttered interface. Despite dominance in the search market, criticisms persist regarding data privacy and search quality.