150 Duplicate page error
-
I am told that I have 150 duplicate page content. It seems that it is the login link on each of my pages. Is this an error? Is it something I have to change?
Thanks
Login/Register at
http://irishdancingdress.com/wp-login.php?redirect_to=http%3A%2F%2Firishdancingdress.com%2Fdress
-
Do you use rel=canonical meta tag? I think if you use it, it will solve your problem.
-
Rel Canonical may not be what you need here.
First question you need to ask yourself is the login page something that needs indexed by Search Engines? If the answer is no, block it with your robots.txt then use -> rel="nofollow" on your login links.
If you have a reason for your login page to be indexed then you'll need to use the meta rel-canonical tag to point to the absolute root of the page.. based on your URL I would assume it is "http://irishdancingdress.com/wp-login.php"
Hope that helps
-
Keith,I think the only way to stop Roger and google from indexing those pages is to put them in the robots.txt file
I made some things global, but Roger seemed to ignore those, so I gave him his own section.
Just modify these to suit your setup.
User-agent: *
Disallow: /tag/*
Disallow: /wp-login.php*User-agent: rogerbot
Disallow: /tag/*
Disallow: /wp-login.php* -
This one's a bit weird - your main "Login" link is fine - this is happening down in the comments section (under "Leave a Reply") - that login link tags the source page, so that you can return to the post.
In this case, I think I'd actually nofollow that and it's probably fine to block it in Robots.txt. This is where things get really situational, as normally I'd advise against that - see my recent post:
http://www.seomoz.org/blog/logic-meet-google-crawling-to-deindex
In your situation, though, Google only seems to be indexing 2 of those URLs currently, so you can probably cut this off before it becomes a problem. Our crawler is being a bit more aggressive in this situation (and, honestly, these links could pose a problem long-term).
If you had a ton of these pages indexed, I'd agree with Slava and recommend rel-canonical, because Robots.txt is pretty ineffective for de-indexing (plus, nofollow causes the problem in my post).
Sorry, I'm making this clear as mud
I think a nofollow and blocking are fine here, because basically the problem hasn't happened yet - you're trying to prevent future problems. You could also monitor for these URLs in Google's index for a few weeks, using this command:site:irishdancingdress.com/wp-login.php
...if that number stays low (it's currently 2), then you're good to go.