Avoiding "Duplicate Page Title" and "Duplicate Page Content" - Best Practices?
-
Hope you don't create link from visitor's researches, like find-a-recipe.php?course=salad**&q=tomatoes** as you would get penalised !
-
Why do you think this? Is it Part of googles terms of service?
-
I manage websites specialising in holiday rentals, so the search pages are very powerful however I only use these for customer experience. For my seo I create pages based on the areas, types of properties, specific searches i.e. villas in florida etc..
I think when building websites you must always have two outlooks; users & seo
-
I noticed this question is still listed as unanswered. Did you come up with a solution you can share with us, and any information about how well it worked? Or are you still looking for advice? Would be great if you could pop back in with an update. Thanks!
-
Hello Baptiste,
I'm keen to know more about why you believe we would get penalised for this. What, specifically, should we seek to avoid in order to avoid the penalty?
Thanks for your help
David
-
I think Baptiste is referring to Google's preference for not including search results in their search results, as the URL in the example appeared to be a search result.
-
Thanks Keri,
Our current experience is that search results from our site are showing up in Google results, sometimes quite high.
So, I'm reluctant to change anything too drastically - "if it ain't broke, don't fix it". But ... maybe we could get slightly higher rankings if we made some minor alterations?
Is there any 'best practice' guidance I could look at to learn more about this specific issue?
Thanks for your help.
David
-
Personally I believe its best practice to have user friendly urls, rather than search generated ones. Google favours this and so do the users. It may be a lot more work to implement, but in my experience (having a site with a lot of categories and posts) it was well worth it.
-
This post is deleted! -
This is something that I've been working on lately. I've been really successfully avoiding duplicate content by using canonical linking, however this has not solve the duplicate Titles nor the duplicate meta descriptions. If you are using a normal web site (static) to post your content as single pages manually, then your only concern would fall into the search pages.
I've switched 100% to Wordpress Blog platforms because of two reasons.
- Google loves them better
- Easier to control content
I've been very successful avoiding duplicate content except for three areas but I do have the solution to repair them as well and I'm currently taking on this task.
The 3 areas of concern are:
- Duplicate Titles
- Duplicate meta descriptions
- Scrapers snatching my unique content and making them their own.
The 3 solutions are: (wordpress platform)
- Duplicate Titles are furnished by pagination next/previous or page #'s at the bottom of each page.
Although wordpress hasn't included this function within the core of its platform yet, Wordpress SEO by Yoast (plugin) automatically add's the new suggested syntax by Google.
Enter rel="next" and rel="prev"
Now, as it goes with these things, Google has just posted the solution. They've asked to add
rel="next"andrel="prev"to paginated archives, so that they can distinguish them as a series and, quote: Send users to the most relevant page/URL—typically the first page of the series.The above syntax will solve our pagination duplicate titles and search paginations. The plugin also adds tag terms at the end of the title for each page. This makes the Title unique.
-
Now the above also tells Google that page #1 is the canonical Title and meta description for all paginations, therefore your meta description is now accurate and safe. The plugin also has an advanced feature which allows you to provide a different description per page other than what the page actually states. Making this slight change makes all the difference.
-
The next problem is robbery or copy infringing my content. My unique content has been scraped and posted without my permission, however now... we can use another rel= syntax to point the article back to the original owner.
rel=”author” and rel=”me” in WP and other platforms
You can allow people to use your content, however the rel="me" tells search engines who the unique content really belongs to. and the rel="author" points to me as well.
This attribute allows you to tell Google who you are as an author and what articles you write. Google has indicated that they believe the authority of an author may even be weighted more heavily than traditional on page metrics, like page or domain authority. As Matt Cutts stated at SMX West, “The concept is that if an author is trustworthy, why does it matter what site the article appears on?”. Author authority also has implications for the impending Panda 2.2 update, which will affect the sites that steal content from other sites to post on their own. If Google sees the same article on 10 different sites, and 1 of those sites clearly identifies an author, marked up with the "rel=author" attribute, which site do you think Google is going to rank?
This is the extent of my research on the above and so far its working well. I hope the above helps for you too.
Cheers!
-
I agree it is best to get the individual pages indexed. Dont have dynamic pages. Instead come up with categories that make sense, have them indexed.
-
The best way to handle this is via the URL Parameters Setting in Google Webmaster or a robots.txt file.
Google added this functionality to handle the exact issues your'e describing, so there's no need to drastically change functionalities which would likely require editing core files in your CMS.
If you click on URL Parameters under Site Configuration in Google Webmaster you will find a list of queries and for each one there available options that instruct google as to how to handle these pages.
To do this:
1. Click Edit for the Paramater you'd like to configure (i.e. course, cooking, etc).
2. In the Dropdown Menu, select Yes. Changes, reorders, or narrow page contents.
3. Choose the option that best describes how the parameter affects the page content.
4. Choose how GoogleBot should crawl these pages.
- I usually choose "Let GoogleBot Decide" as it's Google your trying to please ;). I've designed and optimized several eCommerce store with multiple parameters and this option handles the crawling and indexing of these pages correctly 99% of the time. If you still experience Duplicate Content issues after editing these settings, simply choose the Ignore option.
Dynamic websites are very common these days and this tool is designed by Google specifically to handle parameters in the best possible way and allow Google to understand the URL structure of your site.. The "Don't have dynamic URLs solution" isn't a solution at all, as many modern functionalities rely on dynamic URLs, such as layered navigation in Magento or other eCommerce platforms. How do you suggest filtering products by price, size, color, etc without creating dynamic URLs? These functionalities IMPROVE user experience and navigation. The text in the address bar isn't always the important factor when a user is navigating a site.
Don't overthink it.
Take advantage of the functionality and only de-index pages that are causing duplicate content problems. If you notice specific dynamic URLs are appearing in SERPs too often then create a 301 redirect from that dynamic URL to a landing page with more user friendly URL.
Hope this helps.
Anthony
-
- Setting canonical tag, you should be already doing this rather it's a problem here or not as outlined in seomoz tools you may be missing out on link juice.
I agree with this guy, but I would like to add, why do you want google to crawl your searchable index? Aren't all of your recipes found on your site already by picking categories from a menu?
-
This post is deleted! -
Take Lonnie's advice. Install Yoast. Use the rel=next tags that the software inserts for you automatically. Yoast will fix it all.
Another WP plugin is called htaccess control. And it is also used for this same purpose... It's a little simpler than the Yoast plugin, and if you already have an SEO plugin you like-- or worst yet-- it is built into your theme, like Thesis...
Just go with htaccess control. It is simple to use and your problems will be solved in minutes.
-
Setting canonical tags would be the way I would go, but make sure you have got good seo on the rest of the site for the recipies etc.
-
Solution:
-
Download "All in one SEO" plugin
-
Go to plugin settings. Check these settings:
Use no inxex for -categories
-tags
-search (!)
-Achieves
That will prevent duplicate content issues if you use Wordpress.
-
-
The best way to fix the issue is address it from the server level - so page level creation and urls.
Link canonical is good, but is really a tier three level fix.
Starting at the root is best. You will want to ensure you have:
1. A logical taxonomy, which is a breakdown of the core topic into sub-categories for classification purposes
2. A logical way to tag categories and entities with meaningful tags, or search based on title, content (tags or keys work well - a programmer should be able to help with this)
3. Rewriting the urls as was mentioned so that any urls exported are always exact urls and not using variables or queries
4. 301 redirect appropriate core query urls to the new urls, and implement internal links to the new urls to reinforce that content and show search engines that it is priority.
5. Continue to run reports regularly and monitor the amount of duplicate content. -
Every site you monitor should have a keyword distribution sheet in excel.
Each line will have the url, meta description, title tag, and h1 tag showing (LEN) character count for meta details. With a further column showing the keywords targeted for each page.
With this you have a way to monitor each page in a more direct visual way and avoid duplication, especially titles and meta descriptions.
-
As this question is rather old but still is marked as "unanswered" I take the liberty to post an answer this late which I hope not only you will benefit from but all other webmasters/SEOs with similar issues.
First of all: Duplicate pages and therefrom duplicate title-tags etc. are of course meant to be taken serious but there are no easy fixes in my opinion and especially not if your design and database set-up is causing large amounts of duplicated content.
But is it a big problem Google-wise? I have my doubts based on conflicting signs and indications given by Google Webmaster Tools and the SERPs in general.
An example I just dug up for you: One of the big players in the field of recipes (allrecipes.com) has +5.000 search results indexed by Google ALL with duplicate title-tags which would bring up all the red flags in the Dashboard.
But based on Googles trends for searches, allrecipes.com are still outperforming their closest competitors. Their search result-urls are unique but all have the same title-tag...so +5.000 duplicate title-tags is probably not really a problem in their perspective.
What to do then?
All though your website seems to have been designed with quite a few potential problems built into its core I would personally hesitate spending a lot of resources on fixing it, especially if your traffic from Google is not taking a bashing.
Eventually your website will be in need for a design re-do and perhaps a change of content management system/database system.
Plan ahead and make sure that you will be able to control this issue in your next version.
It could f. i. be by having all search results appear more like individual pages with individual urls. With a little bit of effort you could make each search result unique with unique title tag and url and thereby bring more traffic to your site.
Best of luck with your efforts
Jens Peter