We have all seen the message in Google Search Console’s index protection report for “Web page listed with out content material.” Google’s Gary Illyes stated whenever you see that, more often than not (not all the time), it’s about “pages which might be blocked by robots.txt.”
The error is outlined in Google’s help document as “Web page listed with out content material: This web page seems within the Google index, however for some cause Google couldn’t learn the content material. Attainable causes are that the web page is likely to be cloaked to Google or the web page is likely to be in a format that Google cannot index. This isn’t a case of robots.txt blocking. Examine the web page, and take a look at the Protection part for particulars.”
Gary Illyes was requested if this error may be attributable to “heavy loading time or time-outs” however Gary stated no. If it was a heavy loading time or time-out subject, you’d seemingly see a delicate 404 discover as an alternative Gary defined. Gary stated “this error is de facto only for pages which might be blocked by robots.txt.”
Listed here are these tweets:
no, we might seemingly simply not used these pages of they day out. perhaps we would report them as soft404, relying on whether or not they day out for Googlebot or rendering.
this error is de facto only for pages which might be blocked by robots.txt
— Gary 鯨理／경리 Illyes (@methode) March 20, 2021
Discussion board dialogue at Twitter.