Media in the Age of Algorithms

Go to the profile of Tim O'Reilly
Tim O'ReillyBlockedUnblockFollowFollowing
Founder and CEO, O'Reilly Media. Watching the alpha geeks, sharing their stories, helping the future unfold.
7 mins ago
A prototype of the de Havilland Comet. An object lesson for today. Source: Wikipedia

Media in the Age of Algorithms

Since Tuesday’s election, there’s been a lot of finger pointing, and many of those fingers are pointing at Facebook, arguing that their newsfeed algorithms played a major role in spreading misinformation and magnifying polarization. Some of the articles are thoughtful in their criticism, others thoughtful in their defense of Facebook, while others are full of the very misinformation and polarization that they hope will get them to the top of everyone’s newsfeed. But all of them seem to me to make a fundamental error in how they are thinking about media in the age of algorithms.

Consider Jessica Lessin’s argument in The Information:

“I am deeply, deeply worried about the calls I am hearing, from journalists and friends, for Facebook to intervene and accept responsibility for ensuring citizens are well-informed and getting a balanced perspective….
Facebook promoting trustworthiness sounds great. Who isn’t in favor of accepting responsibility and ferreting out misinformation? But major moves on Facebook’s part to mediate good information from bad information would put the company in the impossible position of having to determine “truth,” which seems far more objective than it really is. Moreover, it would be bad for society.”

My response: Facebook crossed this river long ago. Once they got into the business of curating the newsfeed rather than simply treating it as a timeline, they put themselves in the position of mediating what people are going to see. They became a gatekeeper and a guide. This is not an impossible position. It’s their job. So they’d better make a priority of being good at it.

But those who argue strongly for Facebook’s responsibility to weed out the good from the bad also get it wrong. For example, on Vox, Timothy B. Lee wrote:

A big issue here is about the way Facebook has staffed its editorial efforts. In a traditional news organization, experienced editorial staff occupy senior roles. In contrast, Facebook has relegated the few editorial decisions it has made to junior staffers. For example, until earlier this year Facebook had a team of 15 to 18 independent contractors who were in charge of writing headlines for Facebook’s “trending news” box.
When Facebook faced accusations that these staffers were suppressing conservative stories, Facebook panicked and laid all of them off, running the trending stories box as an automated feature instead. But that hasn’t worked so well either, as fake news keeps popping up in the trending news box.
The problem here wasn’t that Facebook was employing human editors to evaluate stories and write headlines. The problem was that Facebook’s leadership didn’t treat this as an important part of Facebook’s operations.
If Facebook had an experienced, senior editorial team in place, there’s a lot it could do to steer users toward high-quality, deeply reported news stories and away from superficial, sensationalistic, or outright inaccurate ones.

Lee is right to say that curating the news feed isn’t a job for junior staffers and independent contractors. But he’s wrong that it’s a job for “an experienced, senior editorial team.” It’s a job for the brightest minds on Facebook’s algorithm team!

And Lee is wrong to say that the problem wasn’t that Facebook was employing human editors to evaluate stories and write headlines. That was precisely the problem.

Like drivers following a GPS over a bridge that no longer exists, both Jessica Lessin and Timothy Lee are operating from a out-of-date map of the world. In that old map, algorithms are overseen by humans who intervene to compensate for their mistakes. As Jessica rightly notes, this is a very slippery slope.

Jessica says:

…We shouldn’t let Facebook off the hook for every problem it creates or exacerbates. But we can’t hold it responsible for each of them either. We’re witnessing the effects of a world where the internet has driven the cost of saying whatever you want to whomever you want to zero, as Sam often says. This is an irreversible trend no company can stop, nor should we want them to.

But there is a good existence proof for another approach, one that Facebook has worked long and hard to emulate.

Google has long demonstrated that you can help guide people to better results without preventing anyone’s free speech. Like Facebook, they are faced every day with determining which of a thousand competing voices deserve to be at the top of the list. The original insight that Google was founded on, that a link is a vote, and that links from reputable sources that had been around a long time, were worth more than others, was their initial tool for weeding out the wheat from the chaff.

For the better part of two decades, Google has worked tirelessly to thread the needle between curating an algorithmic feed of a firehose of content that can be created by anyone and simply picking winners and losers. Anyone who wants to understand how a 21st century company tackles the problem of editorial curation in a world of infinite information and limited attention would do well to study the history and best practices of Google’s search quality team, which are well documented and widely shared. Here’s one video from Matt Cutts, former head of the Google web spam team.

Google and Facebook constantly devise and test new algorithms. Yes, there is human judgment involved. But it’s judgment applied to the design of a system, not to a specific result. Designing an effective algorithm for search or the newsfeed has more in common with designing an airplane so it flies, or designing a new airplane so that it can fly faster than the old one, than with deciding where that airplane flies.

Improving the “truth value” of articles doesn’t depend on manual interventions to weed out bad results, as commentators on both sides of this issue seem to think, but on discovering signals that cause good results to float to the top.

The question is how to determine what “good results” means.

In the case of making an airplane fly, the goals are simple — stay aloft, go faster, use less fuel — and design changes can be rigorously tested against the desired outcome. There are many analogous problems in search — finding the best price, or the most authoritative source of information on a topic , or a particular document— and many that are far less rigorous. And when users get right to what they want, the users are happy, and so, generally, are advertisers. Unfortunately, unlike search, where the desires of the users to find an answer and get on with their lives are generally aligned with “give them the best results”, Facebook’s prioritization of “engagement” may be leading them in the wrong direction. What is best for Facebook’s revenue may not be best for users.

Even in the case of physical systems like aerodynamics and flight engineering, there are often hidden assumptions to be tested and corrected. In one famous example that determined the future of the aerospace industry, a radically new understanding of how to deal with metal fatigue was needed. As described by University of Texas professor Michael Marder:

Britain was set to dominate the jet age. In 1952, the de Havillands Comet began commercial service, triumphantly connecting London with the farthest reaches of the Empire. The jet plane was years ahead of any competitor, gorgeous to look at, and set new standards for comfort and quiet in the air. Then things went horribly wrong.
In 1953 a Comet fell out of the sky, and the crash was attributed to bad weather and pilot error….In 1954 a second Comet fell out of clear skies near Rome. The fleet was grounded for two months while repairs were made. Flights then resumed with the declaration, “Although no definite reason for the accident has been established, modifications are being embodied to cover every possibility that imagination has suggested as a likely cause of the disaster. When these modifications are completed and have been satisfactorily flight tested, the Board sees no reason why passenger services should not be resumed.” Four days after these words were written a third Comet fell into the sea out of clear skies near Naples and the fleet was grounded again indefinitely.
...As the Comet accident report was being released in 1955, a little-known military contractor in the northwest corner of the United States was completing its prototype for a civilian jet airplane. Boeing had had little success with civilian craft in the past. The company knew that cracks had brought down the Comet, and they had better understand them before they brought down the Boeing 707.
Boeing brought in a researcher for the summer, Paul Paris, a mechanical engineer who had just finished a Master’s degree and was pursuing graduate studies at Lehigh University.…The view of fracture Paris brought to Boeing was dramatically different from the one that had guided construction of the Comet. Cracks were the centerpiece of the investigation. They could not be eliminated. They were everywhere, permeating the structure, too small to be seen. The structure could not be made perfect, it was inherently flawed, and the goal of engineering design was not to certify the airframe free of cracks but to make it tolerate them. [Emphasis mine.]

The essence of algorithm design is not to eliminate all error, but to make results robust in the face of error. Where de Havillands tried in vain to engineer a plane where the materials were strong enough to resist all cracks and fatigue, Boeing realized that the right approach was to engineer a design that allowed cracks, but kept them from propagating so far that they led to catastrophic failure. That is also Facebook’s challenge.

Facebook’s comment in response to Timothy Lee suggests that they understand the challenge they face.

We value authentic communication, and hear consistently from those who use Facebook that they prefer not to see misinformation. In News Feed we use various signals based on community feedback to determine which posts are likely to contain inaccurate information, and reduce their distribution. In Trending we look at a variety of signals to help make sure the topics being shown are reflective of real-world events, and take additional steps to prevent false or misleading content from appearing.
Despite these efforts we understand there’s so much more we need to do, and that is why it’s important that we keep improving our ability to detect misinformation. We’re committed to continuing to work on this issue and improve the experiences on our platform.

The key question to ask is not whether Facebook should be curating the news feed, but HOW. They clearly have a lot of work to do. I believe they are taking the problem very seriously. I hope they make breakthroughs that don’t force them to choose between their business model and giving better results to their users. If they don’t, I fear that despite their good intentions, the business model will win. Their goal is to find a way for the plane to fly faster, but fly safely.

The answer is not for Facebook to put journalists to work weeding out the good from the bad, but to understand, in the same way that they’ve so successfully divined the features that lead to higher engagement, how to build algorithms that take into account “truth” as well as popularity.

And they need to be asking themselves whether they are de Havilland or Boeing.

in The Columbia Journalism Review, Emily Bell argues that

  • BlockedUnblockFollowFollowing
    Go to the profile of Tim O'Reilly

    Tim O'Reilly

    Founder and CEO, O'Reilly Media. Watching the alpha geeks, sharing their stories, helping the future unfold.

    Previous Post
    Next Post