Vulnerabilities / Threats
1/21/2011
02:58 PM
Connect Directly
LinkedIn
Twitter
Google+
RSS
E-Mail
50%
50%

Google Acknowledges Web Spam Complaints

Low-quality content has some Internet users worried about the relevance of Google search results.

Google on Friday tried to quell grumblings about the quality of its search results.

In recent months, prominent bloggers and tech news sites have noted many instances in which Google searches returned poor results. Programmer Jeff Atwood, who runs the popular Coding Horror blog, characterized the volume of complaints as "deafening" lately.

The issue is Web spam. Google principle engineer Matt Cutts, who runs Google's Web spam team, defines Web spam as "the junk you see in search results when Web sites try to cheat their way into higher positions in search results or otherwise violate search engine quality guidelines."

Web spam is a critical issue for Google, perhaps to the point that it imperils Google's search business. If low-quality content continues to find prominent placement in Google's search results and generates enough revenue -- through Google ads, third-party ads, or direct sales -- to fund further Web spam creation, users will slowly but surely turn to other means of content discovery. Social search is often mentioned as a contender in this scenario, which explains why the rise of Facebook has Google worried.

What makes Web spam particularly pernicious is that it's not as easy to identify as malware. Web spam runs the gamut, from blatant attempts to trick Google with unlawfully copied content and repeated search keywords to low-quality writing produced by so-called "content farms."

Cutts's response to the growing chorus of criticism is simultaneously to deny the accuracy of the complaints and to offer assurance that further steps to stamp out Web spam are being taken. Google's search quality is "better than it has ever been in terms of relevance, freshness and comprehensiveness," he insists, even as he acknowledges there has been "a slight uptick of spam in recent months," which Google's engineers are addressing.

Cutts cites a number of steps Google has taken to beat back Web spam, to identify hacked sites, and to alter its search algorithm to deemphasize low-quality Web sites. And he stresses the fact that being a Google advertising client doesn't buy a better search rank.

Cutts concedes that Google can and should do better, even as he suggests that users' perception of the prevalence of Web spam may be the result of "skyrocketing expectations."

The trouble is that Web spammers are trying to do better too.

If Google is to prevail, it may have to look beyond the security arms race, where stalemates rather than victories seem to be the norm, and forgo some ad revenue in order to starve the content farms that feed from Google's trough.

Comment  | 
Print  | 
More Insights
Register for Dark Reading Newsletters
White Papers
Flash Poll
Current Issue
Cartoon
Video
Slideshows
Twitter Feed
Dark Reading - Bug Report
Bug Report
Enterprise Vulnerabilities
From DHS/US-CERT's National Vulnerability Database
CVE-2014-0103
Published: 2014-07-29
WebAccess in Zarafa before 7.1.10 and WebApp before 1.6 stores credentials in cleartext, which allows local Apache users to obtain sensitive information by reading the PHP session files.

CVE-2014-0475
Published: 2014-07-29
Multiple directory traversal vulnerabilities in GNU C Library (aka glibc or libc6) before 2.20 allow context-dependent attackers to bypass ForceCommand restrictions and possibly have other unspecified impact via a .. (dot dot) in a (1) LC_*, (2) LANG, or other locale environment variable.

CVE-2014-0889
Published: 2014-07-29
Multiple cross-site scripting (XSS) vulnerabilities in IBM Atlas Suite (aka Atlas Policy Suite), as used in Atlas eDiscovery Process Management through 6.0.3, Disposal and Governance Management for IT through 6.0.3, and Global Retention Policy and Schedule Management through 6.0.3, allow remote atta...

CVE-2014-2226
Published: 2014-07-29
Ubiquiti UniFi Controller before 3.2.1 logs the administrative password hash in syslog messages, which allows man-in-the-middle attackers to obtains sensitive information via unspecified vectors.

CVE-2014-3020
Published: 2014-07-29
install.sh in the Embedded WebSphere Application Server (eWAS) 7.0 before FP33 in IBM Tivoli Integrated Portal (TIP) 2.1 and 2.2 sets world-writable permissions for the installRoot directory tree, which allows local users to gain privileges via a Trojan horse program.

Best of the Web
Dark Reading Radio