Why are crawlers not picking up these pages?
-
Hi there,
I've been asked to audit a new subdomain for a travel company. It's all a bit messy, so it's going to take some time to remedy. However, one thing I couldn't understand was the low number of pages appearing in certain crawlers.
The subdomain has many pages. A homepage, category pages then product pages. Unfortunately, tools like Screaming Frog and xml-sitemaps.com are only picking up 19 pages and I can't figure out why. Google has so far indexed around 90 pages - this is by no means all of them, but that's probably because of the new domain and lack of sitemap etc.
After looking at the crawl results, only the homepage and category (continent pages) are showing. So all the product pages are not. for example, tours.statravel.co.uk/trip/Amsterdam_Kings_Day_(Start_London_end_London)-COCCKDM11 is not appearing in the crawl results. After reviewing the source code, I can't see anything that would prevent this page being crawled. Am I missing something?
At the moment, the crawl should be picking up around 400+ product pages, but it's not picking up any.
Thanks
-
Hi,
I would think it is the javascript being used on the pages (google can theoretically render the page as a browser would, screaming frog and other similar tools on the whole cannot). If you visit the homepage with js turned off then you see a pretty empty page with a list of links (region, activity, country) which are the same links that screaming frog is picking up. If you go into one of the search results pages with js turned off, you don't really see much of anything at all. Google is obviously doing a better job of crawling the js content! A solution would be to present the data in a simpler, crawlable format for non js enabled browsers but that is (probably a big) conversation with your developers
