Hacker News new | past | comments | ask | show | jobs | submit login

They used to be superb on detecting duplicated content. They also were extremely good at detecting spam/ham. Nowadays it feels like they don't even care anymore and whatever filters they have are either broken or untrained.



Bring back the panda, I say!

https://en.wikipedia.org/wiki/Google_Panda


It is an arms race.

I wonder how many players on the page generation side? The economics of it must be marginal, I guess


Copycat sites also used to be extremely careful at not appearing copycat sites. Or not duplicating content on the same site. I am surely not alone in recalling the old mantra of not duplicating content.

Copycat sites don't seem to care anymore.

I don't believe there are an overwhelming number for Google et al to deal with as it's often the same names topping search results that such filters can remove through semi-manual user action.

While leads to the conclusion - Google don't care about duplicate content any more.


Were they? I remember having to manually block myself a lot of those copycat wikipedia/stackoverflow sites back in 2011 or 2012 when they had the domain-blocklist option available for users. When the feature was removed, it all came back.

Maybe the problem is just that there are more of those now.


Google removed that option without even trying to spin it as a pro-consumer change. The only problems I can think it brought to Google are clueless users complaining that they can no longer see microsoft.com in their results, and having a negative impact on unethical advertisers.


I had lot of problems with "duplicated content" from sites that published the same content as I did and outranked my site.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: