Gone are the days of creating and developing websites with static pages.
But when it comes to discoverability, most websites on the web rely on search engines. They need to appeal search engines, most notably Google, in order to provide them with enough visibility, which in turn, will give them the traffic they need to conduct their business.
While this can be correct, considering that Google sees the content of a page based on the way it is rendered visible, there can be some elements that cannot be rendered properly, purposefully or not, and make Google to fail to see them. If this happens, SEO can be affected in a bad way.
In an episode with Google’s Martin Splitt, John Mueller, Gary Illyes, and Daniel Waisberg at the Search Off the Record podcast, Splitt said that:
But as far as I know, and as far as I can see, we crawl a page and then put the content into the document in our index, and then we render the page, and then we complete the content from the DOM.
In other words, Google said that there is no need to worry when it comes to SEO, as there’s nothing fundamentally different about it compared to static content.
As far as SEO is concerned, Googlebot doesn’t interact with anything when it crawls web pages.
In this case, when crawling and trying to index a page, Google won't click on any button or anything to close that modal box.
As a result, Google won't see the content of the web page, simply because it couldn't.
Google that cannot see the content other than the modal box, won't understand the page and won't rank it on its search results page.
For those who are unsure, they can use the tool 'Fetch as Google' on Google Search Console to see what Googlebox sees when the crawler crawls their web page.
If the Fetch as Google tool is able to render all critical content of a web page like it should, then indeed, there is nothing webmasters and web owners should worry about.