How google bot see's two the same rel canonicals?
-
Hi,
I have a website where all the original URL's have a rel canonical back to themselves. This is kinda like a fail safe mode. It is because if a parameter occurs, then the URL with the parameter will have a canonical back to the original URL.
For example this url: https://www.example.com/something/page/1/ has this canonical: https://www.example.com/something/page/1/ which is the same since it's an original URL
This url https://www.example.com/something/page/1/?parameter has this canonical https://www.example.com/something/page/1/ like i said before, parameters have a rel canonical back to their original url's.
SO: https://www.example.com/something/page/1/?parameter and this https://www.example.com/something/page/1/ both have the same canonical which is this https://www.example.com/something/page/1/
Im telling you all that because when roger bot tried to crawl my website, it gave back duplicates. This happened because it was reading the canonical (https://www.example.com/something/page/1/) of the original url (https://www.example.com/something/page/1/) and the canonical (https://www.example.com/something/page/1/) of the url with the parameter (https://www.example.com/something/page/1/?parameter) and saw that both were point to the same canonical (https://www.example.com/something/page/1/)...
So, i would like to know if google bot treats canonicals the same way. Because if it does then im full of duplicates

thanks.
-
Hello,
Rogerbot struggles a bit with canonical last I checked. You've the right set up you want to stop parameters it's especially helpful for stopping people rankings pages on your site like /?this-site-sucks!
Always remember Rogerbot of any other services are a guide only to help you not a 100% true resource that will help you rank so use them like a tool not an authority.TL:DR - your set up is all ok!

-
All the parameters are configured to NO URL's in google webmaster tools URL parameters tab. Check the image http://prntscr.com/e9fs91
Its a better setting to do it straight from webmaster tools than disallowing the parameters in robots.txt
Tho, i have a problem with that because google is indexing these parameters even if its configured to NO URL's check my post here: https://moz.com/community/q/web-master-tools-url-parameters
-
I would think that you're going a little over the top with what essentially is the job of a canonical tag. you don't need to block robots going to the pages as the canonical tag will be telling robots that its a duplicate version. if the urls have already been indexed it will take time for them to drop off.
-
This post is deleted! -
Its not about the canonical, its about the crawl optimization. I know that canonical URL saves the situation here, i am working under a fail safe mode in matter of duplicates and i want to believe that the canonical URL implementation is better than good in my website.
I just don't want bot's spending time on pages that have nothing actual to say and are canonicalized to pages that have the important content. That is why i configured the bot to not crawl those parameters in the URL parameters tab in GWT and eventually some time to even drop those results.