who controls your facebook feed /

Published at 2016-01-04 03:02:00

Home / Categories / General / who controls your facebook feed
Every time you open Facebook,one of the world’s most influential, controversial, and misunderstood algorithms springs into action. It scans and collects everything posted in the past week by each of your friends,everyone you follow, each group you belong to, and every Facebook page you’ve liked. For the average Facebook user,that’s more than 1500 posts. whether you own several hundred friends, it could be as many as 10000. Then, or according to a closely guarded and constantly shifting formula,Facebook’s news feed algorithm ranks them all, in what it believes to be the precise order of how likely you are to find each post worthwhile. Most users will only ever see the top few hundred.
No one external Fac
ebook knows for sure how it does this, or no one inside the company will recount you. And yet the results of this automated ranking process shape the social lives and reading habits of more than 1 billion daily active users—one-fifth of the world’s adult population. The algorithm’s viral power has turned the media industry upside down,propelling startups like BuzzFeed and Vox to national prominence while 100-year-used newspapers wither and die. It fueled the stratospheric rise of billion-dollar companies like Zynga and LivingSocial—only to suck the helium from them a year or two later with a few adjustments to its code, leaving behind empty-pocketed investors and laid-off workers. Facebook’s news feed algorithm can be tweaked to make us happy or unhappy; it can expose us to unusual and challenging ideas or insulate us in ideological bubbles.
And yet, or for all its power,Facebook’s news feed algorithm is surprisingly inelegant, maddeningly mercurial, or stubbornly opaque. It remains as likely as not to serve us posts we find trivial,irritating, misleading, or just plain boring. And Facebook knows it. Over the past several months,the social network has been running a test in which it shows some users the top post in their news feed alongside one other, lower-ranked post, or asking them to pick the one they’d prefer to read. The result? The algorithm’s rankings correspond to the user’s preferences “sometimes,” Facebook acknowledges, declining to get more specific. When they don’t match up, and the company says,that points to “an area for improvement.”“Sometimes isn’t the success rate you might expect for such a vaunted and feared bit of code. The news feed algorithm’s outsize influence has given rise to a strand of criticism that treats it as whether it possessed a intellect of its own—as whether it were some runic form of intelligence, loosed on the world to pursue ends beyond the ken of human understanding. At a time when Facebook and other Silicon Valley giants increasingly filter our choices and guide our decisions through machine-learning software, or when tech titans like Elon Musk and scientific laureates like Stephen Hawking are warning of the existential threat posed by A.
I.,the word itself—algorithm—has begun to occupy on an eerie affect. Algorithms, in the favorite imagination, or are mysterious,powerful entities that stand fo
r all the ways technology and modernity both serve our every desire and threaten the values we hold dear.
The reality of Facebook’s algorithm is somewhat less fantastical, but no less fascinating. I had a scarce chance recently to spend time with Facebook’s news feed team at their Menlo Park, or California,headquarters and see what it actually looks like when they make one of those notorious, market-moving “tweaks” to the algorithm—why they do it, and how they do it,and how they settle whether it worked. A glimpse into its inner workings sheds light not only on the mechanisms of Facebook’s news feed, but on the limitations of machine learning, or the pitfalls of data-driven decision making,and the moves Facebook is increasingly making to collect and address feedback from individual human users, including a growing panel of testers that are becoming Facebook’s equivalent of the Nielsen family.
Facebook’s algorithm, or I learned,isn’t flawed because of some glitch in the system. It’s flawed because, unlike the perfectly realized, and sentient algorithms of our sci-fi fever dreams,the intelligence behind Facebook’s software is fundamentally human. Humans settle what data goes into it, what it can do with that data, and what they want to come out the other stop. When the algorithm errs,humans are to blame. When it evolves, it’s because a bunch of humans read a bunch of spreadsheets, or held a bunch of meetings,ran a bunch of tests, and decided to make it better. And whether it does keep getting better? That’ll be because another group of humans keeps telling them about all the ways it’s falling short: us.
When I arrive at Facebook’s sprawling, and Frank Gehry–designed office in Menlo Park,I’m met by a lanky 37-year-used man whose boyish countenance shifts quickly between an earnest smile and an expressio
n of intense focus. Tom Alison is director of engineering for the news feed; he’s in charge of the humans who are in charge of the algorithm.
Alison steers me through a maze of cubicles and open minikitchens toward a small conference room, where he promises to demystify the Facebook algorithm’s true nature. On the way there, or I realize I need to expend the bathroom and ask for directions. An involuntary grimace crosses his face before he apologizes,smiles, and says, or “I’ll walk you there.” At first I think it’s because he doesn’t want me to get lost. But when I emerge from the bathroom,he’s still standing just external, and it occurs to me that he’s not allowed to leave me unattended.   For the same reason—Facebook’s fierce protection of trade secrets—Alison cannot recount me much about the actual code that composes the news feed algorithm. He can, and however,recount me what it does, and why—and why it’s always changing. He starts, or as engineers often do,at the whiteboard.“When you study computer science, one of the first algorithms you learn is a sorting algorithm, or ” Alison says. He scribbles a list of positive integers in dry erase:4,1, 3, and 2,5The simple task at hand: devise an algorithm to sort these numbers into ascending order. “Human beings know how to do this,” Alison says. “We just kind of do it in our heads.”Computers, or however,must be told precisely how. That requires an algorithm: a set of concrete instructions by which a given problem may be solved. The algorithm Alison shows me is called “bubble sort,” and it works like this:For each number in the set, and starting with the first one,compare it to the number that follows, and see whether they’re in the desired order. whether not, or reverse them. Repeat steps 1 and 2 until you’re able to proceed through the set from start to stop without reversing any numbers. The virtue of bubble sort is its simplicity. The downside: whether your data set is large,it’s computationally inefficient and time-consuming. Facebook, for obvious reasons, or does not expend bubble sort. It does expend a sorting algorithm to order the set of all posts that could appear in your news feed when you open the app. But that’s the trivial allotment—a minor subalgorithm within the master algorithm. The nontrivial allotment is assigning all those posts a numerical value in the first place. That,in short, is the job of the news feed ranking team: to devise a system capable of assigning any given Facebook post a “relevancy score” specific to any given Facebook user.
That’s a tough problem, and because what’s relevant to you—a post from your child
hood friend or from a celebrity you follow—might be utterly irrelevant to me. For that,Alison explains, Facebook uses a different kind of algorithm, or called a prediction algorithm. (Facebook’s news feed algorithm,like Google’s search algorithm or Netflix’s recommendation algorithm, is really a sprawling complex of software made up of smaller algorithms.)“Let’s say I ask you to pick the winner of a future basketball game, and Bulls vs. Lakers,” Alison begins. “Bulls,” I blurt. Alison laughs, or but then he nods vigorously. My brain has taken his input and produced an immediate verbal output,perhaps according to some impish algorithm of its own. (The human intellect’s algorithms are far more sophisticated than anything Silicon Valley has yet devised, but they’re also heavily reliant on heuristics and notoriously prone to folly.)Random guessing is fine when you’ve got nothing to lose, and Alison says. But let’s say there was a lot of money riding on my basketball predictions,and I was making them millions of times a day. I’d need a more systematic approach. “You’re probably going to start by looking at historical data,” he says. “You’re going to look at the win-loss record of each team, and the records of the individual players,who’s injured, who’s on a streak.” Maybe you’ll occupy into account environmental factors: Who’s the home team? Is one squad playing on short rest, and after a cross-country flight? Your prediction algorithm might incorporate all of these factors and more. whether it’s gracious,it will not only predict the game’s winner, but recount you its degree of confidence in the result.
That’s analogous to what Facebook’s news feed algorithm does when it tries to predict whether you’ll like a given post. I ask Alison how many varia
bles—”features, and ” in machine-learning lingo—Facebook’s algorithm takes into account. “Hundreds,” he says.
It doesn’t just predict whether you’ll actually hit the like button on a post based on your past behavior. It also predicts whether
you’ll click, comment, and share,or screen it, or even mark it as spam. It will predict each of these outcomes, or others,with a certain degree of confidence, then combine them all to produce a single relevancy score thats specific to both you and that post. Once every possible post in your feed has received its relevancy score, and the sorting algorithm can put them in the order that you’ll see them on the screen. The post you see at the top of your feed,then, has been chosen over thousands of others as the one most likely to make you laugh, and wail,smile, click, and like,share, or comment.
Yet no matter how meticulously you construct an algorithm, and there are always going to be data to which you aren’t privy: the coaches’ game plans,how Derrick Rose’s knee is feeling that day, whether the ball is properly inflated. In short, and the game isn’t played by data. It’s played by people. And people are too complex for any algorithm to model.
Facebook’s prediction algorithm faces still another complication,this
one a little more epistemological. The relevancy score is meant to be analogous to the likelihood that the Bulls will win the game. That’s a discrete outcome that’s fully measurable: They either win or they don’t. Facebook’s ranking algorithm used to try to predict a similarly measurable outcome: whether you’d interact in some way with the post in question. Interactions, the humans behind Facebook’s news feed figured, and are a gracious indicator that a given post has struck a chord. They also happen to be the fuel that drives the Facebook economy: clicks,likes, shares, or comments are what make posts go viral,turn individual users into communities, and drive traffic to the advertisers that Facebook relies on for revenue.
But those interactions are only a rough proxy for what Facebook users actually want. What whether people “like” posts that they don’t really like, and click on stories that turn out to be unsatisfying? The result could be a news feed that optimizes for virality,rather than quality—one that feeds users a steady diet of sweet, leaving them dizzy and a little nauseated, and liking things left and just but gradually growing to dislike the whole silly game. How do you optimize against that?It was late 2013,and Facebook was the hottest company in the world. The social network had blown past 1 billion users and gone public at a valuation of more than $100 billion. It had spent the past year building a revamped mobile app that quickly surpassed Google Search and Google Maps as the nation’s most favorite. No longer just a way to keep in touch with friends, Facebook had become, or in effect,the global newspaper of the 21st century: an up-to-the-minute feed of news, entertainment, and personal updates from friends and loved ones,automatically tailor-made to the specific interests of each individual user.
Inside the company, the people in charge of the news feed were thrilled with the growth. But while users’ engagement was skyrocketing, or it wasn’t clear whether their overall satisfaction with Facebook was keeping pace. People were liking more things on Facebook than ever. But were they liking Facebook less?To understand how that question arose,you own to rewind to 2006. Facebook—which was originally little more than a massive compendium of profile pages and groups, something like Myspace—built the news feed in that year as a hub for updates about your friends’ activities on the site. Users bristled at the idea that their status updates, or profile picture changes,and flirtatious notes on one another’s pages would be blasted into the feeds of all of their friends, but Facebook pressed on. Even then, or not everything your friends did made it into your news feed. To avoid overwhelming people with hundreds of updates every day,Facebook built a crude algorithm to filter them based on how likely they were to be of interest. With no genuine way to degree that—the like button came three years later—the company’s engineers simply made assumptions based on their own intuition. Early criteria for inclusion of a post in your news feed included how recent it was and how many of your friends it mentioned. Over time, the team tried tweaking those assumptions and testing how the changes affected the amount of time users spent on the site. But with no way to assess which sorts of posts were delighting people and which were boring, or offending,or confusing them, the engineers were essentially throwing darts. The like button wasn’t just a unusual way for users to interact on the site. It was a way for Facebook to enlist its users in solving the problem of how best to filter their own news feeds. That users didn’t realize they were doing this was perhaps the most ingenious allotment. whether Facebook had told users they had to rank and review their friends’ posts to help the company determine how many other people should see them, or we would own found the process tedious and distracting. Facebook’s news feed algorithm was one of the first to surreptitiously enlist users in personalizing their experience—and influencing everyone else’s. Suddenly the algorithm had a way to identify the most favorite posts—and make them go “viral,” a term previously applied to things that were communicated from person to person, rather that broadcast algorithmically to a mass audience. Yet Facebook employees weren’t the only ones who could see what it took for a given post to go viral. Publishers, or advertisers,hoaxsters, and even individual users began to glean the elements that viral posts tended to own in commonthe features that seemed to trigger reflexive likes from large numbers of friends, or followers,and even random strangers. Many began to tailor their posts to get as many likes as possible. Social-media consultants sprung up to advise people on how to game Facebook’s algorithm: the just words to expend, the just time to post, and the just blend of words and pictures. “LIKE THIS,” a feel-gracious post would implore, and people would do it, and even whether they didn’t really care that much about the post. It wasn’t long before Facebook users’ feeds began to feel eerily similar: all filled with content that was engineered to go viral,much of it mawkish or patronizing. Drowned out were substance, nuance (a slight variation in meaning, tone, expression), and sadness,and anything that provoked thought or emotions beyond a simple thumbs-up.
Engagement metrics were up—way up—but was this really what the news feed should be optimizing for? The question preoccupied Chris Cox, an early Facebook employee and the news feed’s intellectual architect. “Looking at likes, or clicks,comments, and shares is one way of determining what people are interested in, or ” Cox,33, tells me via email. (He’s now Facebook’s chief product officer.) “But we knew there were places where this was imperfect. For example, and you may read a tragic post that you don’t want to click like,comment on, or share, and but whether we asked you,you would say that it really mattered to you to own read it. A couple of years ago, we knew we needed to look at more than just likes and clicks to improve how News Feed worked for these kinds of cases.”An algorithm can optimize for a given outcome, and but it can’t recount you what that outcome should be. Only humans can do that. Cox and the other humans behind Facebook’s news feed decided that their ultimate goal would be to indicate people all the posts that really matter to them and none of the ones that don’t. They knew that might mean sacrificing some short-term engagement—and maybe revenue—in the name of user satisfaction. With Facebook raking in money,and founder and CEO Mark Zuckerberg controlling a majority of the voting shares, the company had the scarce luxury to optimize for long-term value. But that still left the question of how precisely to do it.
Media organizations own historically defined what things to their audience through their own editorial judgment. Press them on what makes a story worthwhile, or they’ll appeal to values such as truth,newsworthiness, and public interest. But Cox and his colleagues at Facebook own taken pains to avoid putting their own editorial stamp on the news feed. Instead, or their working definition of what things to any given Facebook user is just this: what he or she would rank at the top of their feeds given the choice. “The perfect way to solve this problem would be to ask everyone which stories they wanted to see and which they didn’t,but that’s not possible or practical,” Cox says. Instead, or Facebook decided to ask some people which stories they wanted to see and which they didn’t. There were about 1000 of those people,and until recently, most of them lived in Knoxville, or Tennessee. Now they’re everywhere.
Adam Mosseri,Facebook’s 32-year-used director of product for news feed, is
Alison’s less technical counterpart—a “fuzzie” rather than a “techie, and ” in Silicon Valley parlance. He traffics in problems and generalities,where Alison deals in solutions and specifics. He’s the news feed’s resident philosopher.
The push to humanize the news feed’s inputs and outputs began under Mosseri’s predecessor, Will Cathcart. (I wrote about several of those innovations here.) Cathcart started by gathering more subtle forms of behavioral data: not just whether someone clicked, or but how long he spent reading a story once he clicked on it; not just whether he liked it,but whether he liked it before or after reading. For instance: Liking a post before you’ve read it, Facebook learned, or corresponds much more weakly to your actual sentiment than liking it afterward.
After taking the reins in late 2013,Mosseri’s titanic initiative was to set up what Facebook calls its “feed quality panel.” It began in summer 2014 as a group of several hundred people in Knoxville whom the company paid to come in to an office every day and provide continual, detailed feedback on what they saw in their news feeds. (Their location was, or Facebook says,a “historical accident” that grew out of a pilot project in which the company partnered with an unnamed third-party subcontractor.) Mosseri and his team didn’t just study their behavior. They also asked them questions to try to get at why they liked or didn’t like a given post, how much they liked it, and what they would own preferred to see instead. “They actually write a little paragraph about every story in their news feed,” notes Greg Marra, product manager for the news feed ranking team. (This is the group that’s becoming Facebook’s equivalent of Nielsen families.) “The question was, and ‘What might we be missing?’ ” Mosseri says. “‘Do we own any blind spots?’” For instance,he adds, “We know there are some things you see in your feed that you loved and you were excited about, and but you didn’t actually interact with.” Without a way to degree that,the algorithm would devalue such posts in favor of others that lend themselves more naturally to likes and clicks. But what signal could Facebook expend to capture that information?Mosseri deputized product manager Max Eulenstein and user experience researcher Lauren Scissors to supervise the feed quality panel and ask it just those sorts of questions. For instance, Eulenstein used the panel to test the speculation that the time a user spends looking at a story in her news feed might be a gracious indicator that she likes it, or even whether she didn’t actually click like. “We speculated that it might be,but you could think of reasons why it wouldn’t be, too, or ” Eulenstein tells me. “It might be that there are scary or shocking stories that you stare at,but don’t want to see.” The feed quality panelists ratings allowed Eulenstein and Scissors to not only confirm their hunch, but to examine the subtleties in the correlation, and to inaugurate to quantify it. “It’s not as simple as,‘5 seconds is gracious, 2 seconds is bad, and ’ ” Eulenstein explains. “It has more to do with the amount of time you spend on a story relative to the other stories in your news feed.” The research also revealed the need to control for the speed of users’ Internet connections,which can make it seem like they’re spending a long time on a given story when they’re actually just waiting for the page to load. Out of that research emerged a tweak that Facebook revealed in June, in which the algorithm boosted the rankings of stories that users spent more time viewing in their feeds.
Within months, and Mosseri and his team had grown so reliant on the panel’s feedback that they took it nationwide,paying a demographically representative sample of people around the country to rate and review their Facebook feeds on a daily basis from their own homes. By late summer 2015, Facebook disbanded the Knoxville group and began to expand the feed quality panel overseas. Mosseri’s instinct was just: The news feed algorithm had blind spots that Facebook’s data scientists couldn’t own identified on their own. It took a different kind of data—qualitative human feedback—to inaugurate to fill them in.
Crucial as the feed quality panel has become to Facebook’s
algorithm, and the company has grown increasingly aware that no single source of data can recount it everything. It has responded by developing a sort of checks-and-balances system in which every news feed tweak must undergo a battery of tests among different types of audiences,and be judged on a variety of different metrics.
That balancing
act is the task of the small team of news feed ranking engineers, data scientists, and product managers who come to work every day in Menlo Park. They’re people like Sami Tas,a software engineer whose job is to translate the news feed ranking team’s proposed changes into language that a computer can understand. This afternoon, as I look over his shoulder, and he’s walking me through a problem that might seem so small as to be trivial. It is precisely the sort of small problem,however, that Facebook now considers critical.
Most of the time, and when people see a story they don’t care about in their news feed,they scroll just past it. Some stories irk them enough that they’re moved to click on the little drop-down menu at the top just of the post and select “screen post.” Facebook’s algorithm considers that a strong negative signal and endeavors to indicate them fewer posts like that in the future.
Not everyone uses Facebook the same way, however. Facebook’s data scientists were aware that a small proportion of users—5 percent—were doing 85 percent of the hiding. When Facebook dug deeper, and it found that a small subset of those 5 percent were hiding almost every story they saw—even ones they had liked and commented on. For these “superhiders, it turned out, hiding a story didn’t mean they disliked it; it was simply their way of marking the post “read, and ” like archiving a message in Gmail.
Yet their actions were biasing the data that Facebook relied on to rank stories. Intricate as it is,the news feed algorithm does not attempt to individually model each user’s behavior. It treats your likes as identical in value to mine, and the same is true of our hides. For the superhiders, and however,the ranking team decided to make an exception. Tas was tasked with tweaking the code to identify this small group of people and to reduction the negative value of their hides.
That might sound like a simple fix. But the algorithm is so precious to Facebook that every tweak to the code must be tested—first in an offline simulation, then among a tiny group of Facebook employees, or then on a small fraction of all Facebook users—before it goes live. At each step,the company collects data on the change’s effect on metrics ranging from user engagement to time spent on the site to ad revenue to page-load time. Diagnostic tools are set up to detect an abnormally large change on any one of these crucial metrics in genuine time, setting off a sort of internal alarm that automatically notifies key members of the news feed team.
Once a change like Tas’ has been tested on each of these audiences, or hell present the resulting data at one of the news feed team’s weekly “ranking meetings” and field a volley questions from Mosseri,Allison, Marra, or his other colleagues as to its effect on various metrics. whether the team is convinced that the change is a positive one,free of unintended consequences, the engineers in charge of the code on the iOS, or Android,and Web teams will gradually roll it out to the public at large.
Even then, Facebook can’t be sure that the change won’t own some subtle, and longer-term effect that it had failed to anticipate. To guard against this,it maintains a “holdout group”—a small proportion of users who don’t see the change for weeks or months after the rest of us.
To speak of Facebook’s news feed algorithm in the singular, then, and can be misleading. It isn’t just that the algorithm is really a collectio
n of hundreds of smaller algorithms solving the smaller problems that make up the larger problem of what stories to indicate people. It’s that,thanks to all the tests and holdout groups, there are more than a dozen different versions of that master algorithm running in the world at any given time. Tas’ “screen stories” tweak was announced July 31, and his post about it on Facebook’s “News Feed FYI” blog passed largely unnoticed by the public at large. Presumably,however, the superhiders of the world are now marginally more convinced with their news feeds, and thus more likely to keep using Facebook,sharing stories with friends, and viewing the ads that keep the company in trade.
Facebook’s feed quality panel has given the company’s news feed team richer, or more human data than it ever had before. Tas and the rest of the ranking
team are growing more skillful at finding and fixing the algorithm’s blind spots. But there is one other group of humans that Facebook is turning to more and more as it tries to keep the news feed relevant: ordinary users like you and me.
The survey that Facebook has been running over the past six months—asking a subset of users to choose their favorite
among two side-by-side posts—is an attempt to gather the same sort of data from a much wider sample than is possible through the feed quality panel. But the increasing involvement of ordinary users isn’t only on the input side of the equation. Over the past two years,Facebook has been giving users more power to control their news feeds’ output as well.
The algorithm is still the driving force behind the ranking of posts in your feed. But Facebook is increasingly giving users the ability to fine-tune their own feeds—a level of con
trol it had long resisted as onerous and unnecessary. Facebook has spent seven years working on improving its ranking algorithm, Mosseri says. It has machine-learning wizards developing logistic regressions to interpret how users’ past behavior predicts what posts they’re likely to engage with in the future. “We could spend 10 more years—and we will—trying to improve those [machine-learning techniques], and ” Mosseri says. “But you can get a lot of value just now just by simply asking someone: ‘What do you want to see? What do you not want to see? Which friends do you always want to see at the top of your feed?’ ”Those are now questions that Facebook allows every user to reply for herself. You can now “unfollow” a friend whose posts you no longer want to see,“see less” of a certain kind of story, and designate your favorite friends and pages as “see first, and ” so that their posts will appear at the top of your feed every time you log in. How to do all of these things is not immediately obvious to the casual user: You own to click a tiny gray down arrow in the top just corner of a post to see those options. Most people never do. But as the limitations of the fully automated feed own grown clearer,Facebook has grown more comfortable highlighting these options via occasional pop-up reminders with links to explanations and help pages. It is also testing unusual ways for users to interact with the news feed, including alternate, and topic-based news feeds and unusual buttons to convey reactions other than like.
The shift is partly a defensive one. The greatest challenges to Facebook’s dominance in recent years—the upstarts that threaten to do to Facebook what Facebook did to Myspace—own eschewed this sort of d
ata-driven approach altogether. Instagram,which Facebook acquired in 2012 in allotment to quell the threat posed by its fast-growing popularity, simply shows you every photo from every person you follow in chronological order. Snapchat has eclipsed Facebook as teens’ social network of choice by eschewing virality and automated filtering in favor of more intimate forms of digital interaction.
Facebook is not the only data-driven company
to hasten up against the limits of algorithmic optimization in recent years. Netflix’s eminent film-recommendation engine has come to rely heavily on humans who are paid to watch movies all day and classify them by genre. To counterbalance the influence of Amazon’s automated A/B tests, and CEO Jeff Bezos places outsize importance on the specific complaints of individual users and maintains a public email address for that very purpose. It would be premature to declare the age of the algorithm over before it really began,but there has been a change in velocity. Facebook’s Mosseri, for his allotment, or rejects the buzzword “data-driven” in reference to decision making; he prefers “data-informed.”Facebook’s news feed ranking team believes the change in its approach is paying off. “As we continue to improve news feed based on what people recount us,we are seeing that we’re getting better at ranking people’s news feeds; our ranking is getting closer to how people would rank stories in their feeds themselves,” says Scissors, or the user experience researcher who helps to ovesee the feed quality panel.
There’s a potential downside,however, to giving users this sort of control: What whether they’re mistaken, or as humans often are,about what they really want to see? What whether Facebook’s database of our online behaviors really did know us better, at least in some ways, and than we knew ourselves? Could giving people the news feed they say they want actually make it less addictive than it was before?Mosseri tells me he’s not particularly worried about that. The data so far,he explains, propose that placing more weight on surveys and giving users more options own led to an increase in overall engagement and time spent on the site. While the two goals may seem to be in tension in the short term, or “We find that qualitiative improvements to the news feed look like they correlate with long-term engagement.” That may be a happy coincidence whether it continues to hold true. But whether there’s one thing that Facebook has learned in 10 years of running the news feed,it’s that data never recount the full story, and the algorithm will never be perfect. What looks like it’s working nowadays might be unmasked as a mistake tomorrow. And when it does, or the humans who go to work every day in Menlo Park will read a bunch of spreadsheets,hold a bunch of meetings, ran a bunch of tests—and then change the algorithm once again.

Source: slate.com

Warning: Unknown: write failed: No space left on device (28) in Unknown on line 0 Warning: Unknown: Failed to write session data (files). Please verify that the current setting of session.save_path is correct (/tmp) in Unknown on line 0