Three weeks ago I add this in my robots.txt to disallow pagination so it will not appear in the Google index:
User-Agent: *
Disallow: /blog/1
Disallow: /blog/2
Disallow: /blog/3
Disallow: /blog/4
Disallow: /blog/5
Disallow: /blog/6
Disallow: /blog/7
Disallow: /blog/8
Disallow: /blog/9
When Google was allowed to crawl these pagination pages, they gave result like this when I searched a keyword in my blog:
Is my robots.txt line above I intend to disallow only the pagination, but not the articles. The crawler should be able to still index the article.
Today Ahrefs reported around 173K uncrawled pages. Is it caused by my new robots.txt disallows? I suspect it is either that or because after migration there things still need time to settle.
There is this message too:
The website wasn't fully crawled
The crawl has reached the maximum depth level from the seed, and the website has not been crawled completely.
To crawl deeper levels of your site, increase the "Max depth level from seed" in the project settings and start a new crawl.
You might also want to check why some of the URLs on your website are considerably distant from the seed.
Can someone explain it to me about this? I'm confused since it is the first time I've stumbled upon this problem.
