Vulnerabilities / Threats
1/21/2011
02:58 PM
Connect Directly
LinkedIn
Twitter
Google+
RSS
E-Mail
50%
50%

Google Acknowledges Web Spam Complaints

Low-quality content has some Internet users worried about the relevance of Google search results.

Google on Friday tried to quell grumblings about the quality of its search results.

In recent months, prominent bloggers and tech news sites have noted many instances in which Google searches returned poor results. Programmer Jeff Atwood, who runs the popular Coding Horror blog, characterized the volume of complaints as "deafening" lately.

The issue is Web spam. Google principle engineer Matt Cutts, who runs Google's Web spam team, defines Web spam as "the junk you see in search results when Web sites try to cheat their way into higher positions in search results or otherwise violate search engine quality guidelines."

Web spam is a critical issue for Google, perhaps to the point that it imperils Google's search business. If low-quality content continues to find prominent placement in Google's search results and generates enough revenue -- through Google ads, third-party ads, or direct sales -- to fund further Web spam creation, users will slowly but surely turn to other means of content discovery. Social search is often mentioned as a contender in this scenario, which explains why the rise of Facebook has Google worried.

What makes Web spam particularly pernicious is that it's not as easy to identify as malware. Web spam runs the gamut, from blatant attempts to trick Google with unlawfully copied content and repeated search keywords to low-quality writing produced by so-called "content farms."

Cutts's response to the growing chorus of criticism is simultaneously to deny the accuracy of the complaints and to offer assurance that further steps to stamp out Web spam are being taken. Google's search quality is "better than it has ever been in terms of relevance, freshness and comprehensiveness," he insists, even as he acknowledges there has been "a slight uptick of spam in recent months," which Google's engineers are addressing.

Cutts cites a number of steps Google has taken to beat back Web spam, to identify hacked sites, and to alter its search algorithm to deemphasize low-quality Web sites. And he stresses the fact that being a Google advertising client doesn't buy a better search rank.

Cutts concedes that Google can and should do better, even as he suggests that users' perception of the prevalence of Web spam may be the result of "skyrocketing expectations."

The trouble is that Web spammers are trying to do better too.

If Google is to prevail, it may have to look beyond the security arms race, where stalemates rather than victories seem to be the norm, and forgo some ad revenue in order to starve the content farms that feed from Google's trough.

Comment  | 
Print  | 
More Insights
Register for Dark Reading Newsletters
White Papers
Cartoon
Current Issue
Dark Reading, September 16, 2014
Malicious software is morphing to be more targeted, stealthy, and destructive. Are you prepared to stop it?
Flash Poll
Video
Slideshows
Twitter Feed
Dark Reading - Bug Report
Bug Report
Enterprise Vulnerabilities
From DHS/US-CERT's National Vulnerability Database
CVE-2014-3090
Published: 2014-09-23
IBM Rational ClearCase 7.1 before 7.1.2.15, 8.0.0 before 8.0.0.12, and 8.0.1 before 8.0.1.5 allows remote attackers to cause a denial of service (memory consumption) via a crafted XML document containing a large number of nested entity references, a similar issue to CVE-2003-1564.

CVE-2014-3101
Published: 2014-09-23
The login form in the Web component in IBM Rational ClearQuest 7.1 before 7.1.2.15, 8.0.0 before 8.0.0.12, and 8.0.1 before 8.0.1.5 does not insert a delay after a failed authentication attempt, which makes it easier for remote attackers to obtain access via a brute-force attack.

CVE-2014-3103
Published: 2014-09-23
The Web component in IBM Rational ClearQuest 7.1 before 7.1.2.15, 8.0.0 before 8.0.0.12, and 8.0.1 before 8.0.1.5 does not set the secure flag for the session cookie in an https session, which makes it easier for remote attackers to capture this cookie by intercepting its transmission within an http...

CVE-2014-3104
Published: 2014-09-23
IBM Rational ClearQuest 7.1 before 7.1.2.15, 8.0.0 before 8.0.0.12, and 8.0.1 before 8.0.1.5 allows remote attackers to cause a denial of service (memory consumption) via a crafted XML document containing a large number of nested entity references, a similar issue to CVE-2003-1564.

CVE-2014-3105
Published: 2014-09-23
The OSLC integration feature in the Web component in IBM Rational ClearQuest 7.1 before 7.1.2.15, 8.0.0 before 8.0.0.12, and 8.0.1 before 8.0.1.5 provides different error messages for failed login attempts depending on whether the username exists, which allows remote attackers to enumerate account n...

Best of the Web
Dark Reading Radio