In Weapons of Math Destruction, author Cathy O’Neil writes that humanity is in the midst of a “new revolution.” “Big Data”—a field that uses huge swaths of data to make various industries more efficient or profitable—is rapidly changing the way society functions. And while O’Neil acknowledges that data collection and computer algorithms can be helpful in certain contexts, she also warns that behind the scenes, much of modern life is dictated by machines rather than people. Companies are increasingly using computer-based algorithms to interpret data and make important decisions (like who gets interviewed for a job or who can secure a loan). Moreover, computers are controlling our lives in potentially unfair and inhumane ways, since the algorithms they depend on can be inaccurate, biased, or otherwise flawed—and O’Neil suggests that this will continue unless humans play a more active role in data interpretation. In order to keep technology’s influence over humans in check, corporations must implement internal regulations on how data is gathered, employ data scientists to make sure it’s interpreted correctly, and ensure that developers are held accountable for the effects of the algorithms they create.
In recent decades, more and more sectors of the economy have begun to gather and use data in increasingly sophisticated ways—and some of this data has the potential to improve human lives. For instance, in the mid-1980s, U.S. News & World Report started releasing data-backed college rankings. Many colleges that were dissatisfied with their rankings took steps to improve their schools by fundraising, admitting a more diverse student body, and building better infrastructures for their campuses. Another example is standardized tests, which gather data about students’ performances in American public schools. These tests have the potential to help teachers tailor their curriculum to their students’ needs and to direct more funding to school districts that need additional resources. Lastly, O’Neil uses the example of trucking companies that have begun to more closely track and surveil their truckers’ rigs. By installing cameras and GPS devices and monitoring how truckers are driving at different hours of the day, they can gather data about when their drivers might be struggling to stay awake—and thus prevent tragic or fatal accidents.
But although data collection can benefit humanity, relying too much on computer algorithms poses an existential and moral problem. In the book’s conclusion, O’Neil suggests that algorithms and mathematical models that promise to make life more efficient by erasing human bias and error also end up erasing things that only humans can do: imagine, invent, and self-correct. “Compared to the human brain, machine learning isn’t especially efficient,” O’Neil writes. Machines can’t differentiate between the truth and lies—they can only analyze data. This means that mathematical models and programs may be encoded with human bias—they are created by humans, after all. And if biased or otherwise faulty algorithms run on their own without being vetted or regulated by humans along the way, they may make connections based on flawed associations. As “error-ridden” as these computer systems may be, they play a huge role in determining some of the most important parts of modern-day life. For instance, many colleges use algorithms to screen applicants’ personal information, and financial institutions use algorithms to decide who can or can’t secure credit. Algorithms have an enormous amount of control over human lives—and yet they’re not regulated or held to any kind of standard.
O’Neil argues that to make sure these programs and machines stay on track, direct human involvement at every stage of an algorithm’s development and implementation is necessary. First, O’Neil suggests, data should be gathered in more transparent ways, so that the public understands when and why their personal information is being used. Second, the creators of algorithms and models need to take greater responsibility for their creations—O’Neil suggests that they should be made to take an oath similar to the Hippocratic Oath doctors take before practicing medicine, whose most famous promise is to “do no harm.” Lastly, O’Neil suggests that in-house data scientists (as well as external advisors) need to vet these algorithms before they’re put to use, and make sure that they’re gathering and using data fairly. Together, all of these steps would help ensure that technology doesn’t have undue control over humanity, and that the algorithms various industries use are helpful rather than harmful. It’s up to humanity, O’Neil suggests, to recognize that we’re not living in a “techno-utopia.” Instead, we’re at a very fragile moment in history that needs to be navigated delicately in order to put a stop to the flawed, harmful algorithms that O’Neil calls “weapons of math destruction.”
Humanity vs. Technology ThemeTracker
Humanity vs. Technology Quotes in Weapons of Math Destruction
The math-powered applications powering the data economy were based on choices made by fallible human beings. Some of these choices were no doubt made with the best intentions. Nevertheless, many of these models encoded human prejudice, misunderstanding, and bias into the software systems that increasingly managed our lives. Like gods, these mathematical models were opaque, their workings invisible to all but the highest priests in their domain: mathematicians and computer scientists. Their verdicts, even when wrong or harmful, were beyond dispute or appeal. And they tended to punish the poor and the oppressed in our society, while making the rich richer.
Do you see the paradox? An algorithm processes a slew of statistics and comes up with a probability that a certain person might be a bad hire, a risky borrower, a terrorist, or a miserable teacher. That probability is distilled into a score, which can turn someone’s life upside down. And yet when the person fights back, “suggestive” countervailing evidence simply won’t cut it. The case must be ironclad. The human victims of WMDs, we’ll see time and again, are held to a far higher standard of evidence than the algorithms themselves.
The value-added model in Washington, D.C., schools […] evaluates teachers largely on the basis of students’ test scores, while ignoring how much the teachers engage the students, work on specific skills, deal with classroom management, or help students with personal and family problems. It’s overly simple, sacrificing accuracy and insight for efficiency. Yet from the administrators’ perspective it provides an effective tool to ferret out hundreds of apparently underperforming teachers, even at the risk of misreading some of them.
And here’s one more thing about algorithms: they can leap from one field to the next, and they often do. Research in epidemiology can hold insights for box office predictions; spam filters are being retooled to identify the AIDS virus. This is true of WMDs as well. So if mathematical models in prisons appear to succeed at their job—which really boils down to efficient management of people—they could spread into the rest of the economy along with the other WMDs, leaving us as collateral damage.
That’s my point. This menace is rising.
Paradoxically, the supposedly powerful algorithms that created the market, the ones that analyzed the risk in tranches of debt and sorted them into securities, turned out to be useless when it came time to clean up the mess and calculate what all the paper was actually worth. The math could multiply the horseshit, but it could not decipher it. This was a job for human beings. Only people could sift through the mortgages, picking out the false promises and wishful thinking and putting real dollar values on the loans.
What does a single national diet have to do with WMDs? Scale. A formula, whether it’s a diet or a tax code, might be perfectly innocuous in theory. But if it grows to become a national or global standard, it creates its own distorted and dystopian economy. This is what has happened in higher education.
It sounds like a joke, but they were absolutely serious. The stakes for the students were sky high. As they saw it, they faced a chance either to pursue an elite education and a prosperous career or to stay stuck in their provincial city, a relative backwater. And whether or not it was the case, they had the perception that others were cheating. So preventing the students in Zhongxiang from cheating was unfair. In a system in which cheating is the norm, following the rules amounts to a handicap.
The Internet provides advertisers with the greatest laboratory ever for consumer research and lead generation. […] Within hours […], each campaign can zero in on the most effective messages and come closer to reaching the glittering promise of all advertising: to reach a prospect at the right time, and with precisely the best message to trigger a decision, and thus succeed in hauling in another paying customer. This fine-tuning never stops.
And increasingly, the data-crunching machines are sifting through our data on their own, searching for our habits and hopes, fears and desires.
For-profit colleges, sadly, are hardly alone in deploying predatory ads. They have plenty of company. If you just think about where people are hurting, or desperate, you’ll find advertisers wielding their predatory models.
These types of low-level crimes populate their models with more and more dots, and the models send the cops back to the same neighborhood.
This creates a pernicious feedback loop. The policing itself spawns new data, which justifies more policing. And our prisons fill up with hundreds of thousands of people found guilty of victimless crimes. Most of them come from impoverished neighborhoods, and most are black or Hispanic. So even if a model is color blind, the result of it is anything but. In our largely segregated cities, geography is a highly effective proxy for race.
Police make choices about where they direct their attention. Today they focus almost exclusively on the poor. […] And now data scientists are stitching this status quo of the social order into models, like PredPol, that hold ever-greater sway over our lives.
The result is that while PredPol delivers a perfectly useful and even high-minded software tool, it is also a do-it-yourself WMD. In this sense, PredPol, even with the best of intentions, empowers police departments to zero in on the poor, stopping more of them, arresting a portion of those, and sending a subgroup to prison. […]
The result is that we criminalize poverty, believing all the while that our tools are not only scientific but fair.
While looking at WMDs, we’re often faced with a choice between fairness and efficacy. Our legal traditions lean strongly toward fairness. The Constitution, for example, presumes innocence and is engineered to value it. […]
WMDs, by contrast, tend to favor efficiency. By their very nature, they feed on data that can be measured and counted. But fairness is squishy and hard to quantify. It is a concept.
The hiring business is automating, and many of the new programs include personality tests like the one Kyle Behm took. It is now a $500 million annual business and is growing by 10 to 15 percent a year […]. Such tests now are used on 60 to 70 percent of prospective workers in the United States […].
Naturally, these hiring programs can't incorporate information about how the candidate would actually perform at the company. That’s in the future, and therefore unknown. So like many other Big Data programs, they settle for proxies. And as we’ve seen, proxies are bound to be inexact and often unfair.
The key is to analyze the skills each candidate brings […], not to fudge him or her by comparison with people who seem similar. What’s more, a bit of creative thinking at St. George’s could have addressed the challenges facing women and foreigners. […]
This is a point I’ll be returning to in future chapters: we’ve seen time and again that mathematical models can sift through data to locate people who are likely to face great challenges, whether from crime, poverty, or education. It’s up to society whether to use that intelligence to reject and punish them—or to reach out to them with the resources they need. We can use the scale and efficiency that make WMDs so pernicious in order to help people.
Phrenology was a model that relied on pseudoscientific nonsense to make authoritative pronouncements, and for decades it went untested. Big Data can fall into the same trap. Models like the ones that red-lighted Kyle Behm and blackballed foreign medical students at St. George’s can lock people out, even when the “science” inside them is little more than a bundle of untested assumptions.
With Big Data, […] businesses can now analyze customer traffic to calculate exactly how many employees they will need each hour of the day. The goal, of course, is to spend as little money as possible, which means keeping staffing at the bare minimum while making sure that reinforcements are on hand for the busy times.
But data studies that track employees’ behavior can also be used to cull a workforce. As the 2008 recession ripped through the economy, HR officials in the tech sector started to look at those Cataphora charts with a new purpose. They saw that some workers were represented as big dark circles, while others were smaller and dimmer. If they had to lay off workers, and most companies did, it made sense to start with the small and dim ones on the chart.
Were those workers really expendable? Again we come to digital phrenology. If a system designates a worker as a low idea generator or weak connector, that verdict becomes its own truth. That’s her score.
While its scores are meaningless, the impact of value-added modeling is pervasive and nefarious. “I’ve seen some great teachers convince themselves that they were mediocre at best based on those scores,” Clifford said. “It moved them away from the great lessons they used to teach, toward increasing test prep. To a young teacher, a poor value-added score is punishing, and a good one may lead to a false sense of accomplishment that has not been earned.”
Since [the invention of the FICO score], the use of scoring has of course proliferated wildly. Today we’re added up in every conceivable way as statisticians and mathematicians patch together a mishmash of data, from our zip codes and Internet surfing patterns to our recent purchases. Many of their pseudoscientific models attempt to predict our creditworthiness, giving each of us so-called e-scores. These numbers, which we rarely see, open doors for some of us, while slamming them in the face of others. Unlike the FICO scores they resemble, e-scores are arbitrary, unaccountable, unregulated, and often unfair—in short, they’re WMDs.
Big Data processes codify the past. They do not invent the future. Doing that requires moral imagination, and that's something only humans can provide. We have to explicitly embed better values into our algorithms, creating Big Data models that follow our ethical lead. Sometimes that will mean putting fairness ahead of profit.
Data is not going away. […] Predictive models are, increasingly, the tools we will be relying on to run our institutions, deploy our resources, and manage our lives. But as I’ve tried to show throughout this book, these models are constructed not just from data but from the choices we make about which data to pay attention to—and which to leave out. Those choices are not just about logistics, profits, and efficiency. They are fundamentally moral.
If we back away from them and treat mathematical models as a neutral and inevitable force […] we abdicate our responsibility. And the result, as we’ve seen, is WMDs that treat us like machine parts […] and feast on inequities. We must come together to police these WMDs, to tame and disarm them.