Lately I have been seeing a lot of websites that aggregate content from around the web (especially from social media) to compile in a ‘mashup’ style format. Of course, this has been happening for years, but only just started reaching new heights in popularity. With the success of readers like Feeddler and Pulse, it is only natural that other sites started to take advantage of platforms that let them fill that need.
But there is a nagging question faced by those who have adopted this practice. Will Google’s bots read it as duplicate content? Is there a way to make it so they don’t?
How Bots See It
The truth is, this is a very hard thing to know for sure. Because the pages that the bots check are random, it is impossible to know what they will be looking at. So if you have some pages that are not as mixed, there is a good chance they will read it and other pages as duplicated from content around the web.
While there is technically no penalty for this, it will banish those pages from the main results. Which is a punishment in and of itself. After all, how will anyone find it?
Only through making sure all content you mashup on your site is meaningful, and mixed enough not to be identical to any other website, will you be more likely not to look like duplicate content. But even then, if it violates the terms of the site you took it from, you are in for a headache. Not to mention a take down notice.
Getting Proper SEO Results
If you are concerned about bringing in good results, there is only one way to do it: providing unique, high quality content. Which cannot be done through purely aggregated links and references from other sites. You might not get struck with duplicate content, but it doesn’t mean you will be drawing any attention to yourself, or helping your rankings/traffic.
You have to make sure that you are writing unique content, while mixing up aggregated content enough to classify as unique content. Otherwise, you won’t be doing yourself any favors.