Why are only PDFs on my client's site being indexed, and not actual pages?
-
My client has recently built a new site (we did not build this), which is a subdomain of their main site. The new site is: https://addstore.itelligencegroup.com/uk/en/. (Their main domain is: http://itelligencegroup.com/uk/)
This new Addstore site has recently gone live (in the past week or so) and so far, Google appears to have indexed 56 pdf files that are on the site, but it hasn't indexed any of the actual web pages yet. I can't figure out why though.
I've checked the robots.txt file for the site which appears to be fine: https://addstore.itelligencegroup.com/robots.txt.
Does anyone have any ideas about this?
-
Did you try FETCH AS GOOGLE web master tool to index your sub domain pages?
-
Hi Ramansaab
Yes, but I have asked the client to change the way that Webmasters is set up because at the moment when we try to submit a page, it points to a redirect. So because Webmasters can't follow a redirect I have to wait for the client to set it up with the right url for the domain. They have set it up as www.addstore.itelligencegroup.com/uk/en when in fact it should be addstore.itelligencegroup.com/uk/en.
-
Hi Maureen!
It seems that your pages all have the following code inside the :
<meta content="<a class="attribute-value">noindex,follow</a>" name="<a class="attribute-value">robots</a>">
It will effectively block all your pages from being indexed, and should be removed (or changed into <meta content="<a class="attribute-value">index,follow</a>" name="<a class="attribute-value">robots</a>"> )Also, you should update the robots.txt to reference the full URL for the XML sitemap? I think the specifications for referenging this demands full URL's.
Hope this helps

Anders -
Thank you so much, Anders, I am getting the client to change this immediately. Brilliant answer!