Content Moderation Case Study: Dealing With Misinformation In Search (2004)
Furnished content.
This series of case studies is published in partnership with the Trust & Safety Foundation to examine the difficult choices and tradeoffs involved in content moderation. Learn more »Summary: Google's biggest early innovation in search was that it used inbound links as a tool for determining the popularity of a website, and thus what its relevance to a particular search might be. That feature, however, created some side effects that raised concerns about how search results might lead to misinformation, or how the search engine might be gamed.One of the earliest examples of this was the discovery in 2004 that the first result of a search on the word jew pointed to a blatantly anti-semitic website, Jewwatch. It was widely theorized that the reason for this was that the singular noun jew was more likely to be used by those pushing anti-semitic arguments, rather than the more common adjective jewish or the phrase jewish wo/man etc. Also, the site Jewwatch had been in existence for many years, and had many inbound links from other sources.Some also believed that the people behind Jewwatch had used an early search engine optimization technique known as Googlebombing to purposefully game the results deliberately linking to Jewwatch from other sites, and using the word jew as the link text.As this result got attention, Google came under tremendous pressure to change the search result, as people accused the company of anti-semitism or deliberately pointing to the Jewwatch site in search results. The Anti-Defamation League sent a letter to Google asking it to explore whether or not its ranking system needed to be changed (though the ADL also posted an article to its own site telling people that it was clear that the result was not intentional, or done for nefarious reasons). Some politicians, including Senator Chuck Schumer, also got involved to pressure Google to change its results.Decisions to be made by Google:
People have asked about how we feel about Googlebombs, and we have talked about them in the past. Because these pranks are normally for phrases that are well off the beaten path, they haven't been a very high priority for us. But over time, we've seen more people assume that they are Google's opinion, or that Google has hand-coded the results for these Googlebombed queries. That's not true, and it seemed like it was worth trying to correct that misperception. So a few of us who work here got together and came up with an algorithm that minimizes the impact of many Googlebombs.The next natural question to ask is "Why doesn't Google just edit these search results by hand?" To answer that, you need to know a little bit about how Google works. When we're faced with a bad search result or a relevance problem, our first instinct is to look for an automatic way to solve the problem instead of trying to fix a particular search by hand. Algorithms are great because they scale well: computers can process lots of data very fast, and robust algorithms often work well in many different languages. That's what we did in this case, and the extra effort to find a good algorithm helps detect Googlebombs in many different languages. We wouldn't claim that this change handles every prank that someone has attempted. But if you are aware of other potential Googlebombs, we are happy to hear feedback in our Google Web Search Help Group.
edit: Policy/auto___content_moderation_case_study__dealing_with_misinformation_in_search__2004_.wikieditish...