Discussion about this post

User's avatar
LeMain's avatar

Fascinating insight regarding the 'Priority Crawler' vs. the 'Regular Crawler.' The point that we essentially get one chance to optimize a news article before it drops out of the immediate crawl cycle is a crucial reminder for editorial workflows.

It really highlights why understanding the fundamental difference between crawling and indexing is so important before diving into these advanced news SEO strategies. I recently broke down the core mechanics of how Google discovers and indexes pages for those who are still mastering the basics, For a newly created website, Googlebot rarely initiates crawls automatically—it often doesn't even know the site exists yet. Therefore, proactive submission in the early stages is absolutely necessary, this is just a supplementary note regarding the phase before Googlebot starts crawling proactively: https://www.cybrinal.com/blog/what-is-google-index-how-to-get-site-indexed/

Thanks for the deep dive on the robots.txt nuances as well!

Expand full comment
Matko's avatar

Question; what if in Search console we have - for the same url - "Crawl request: JSON" 200, and "Page indexing" 404 response ? Lets say its some lazy loaded element. Should crawl to those url's be blocked trough robots.txt ?

Expand full comment

No posts

Ready for more?