Machine See, Machine Do: How Technology Mirrors Bias in Our Criminal Justice System
()
About this ebook
"When today's technology relies on yesterday's data, it will simply mirror our past mistakes and biases."
AI and other high-tech tools embed and reinforce America's history of prejudice and exclusion - even when they are used with the best intentions. Patrick K. Lin's Machine See, Machine Do: How Technology Mirrors Bias in
Patrick K. Lin
??????? ?. ??? is an attorney and researcher focused on AI, privacy, and technology regulation. He is the author of ??????? ???, ??????? ??, a book that explores the ways public institutions use technology to surveil, police, and make decisions about the public, as well as the historical biases that impact that technology.Patrick has extensive experience in litigation and policy, having worked for the ACLU, FTC, EFF, and other organizations that advocate for digital rights and social justice. He is passionate about addressing the ethical and legal challenges posed by emerging technologies, especially in the areas of surveillance, algorithmic bias, and data privacy. He serves as the junior board chair of the Surveillance Technology Oversight Project (STOP). He has also published multiple articles and papers on topics such as facial recognition, data protection, and copyright law.
Related to Machine See, Machine Do
Related ebooks
Speech Police: The Global Struggle to Govern the Internet Rating: 0 out of 5 stars0 ratingsJustice by Means of Democracy Rating: 0 out of 5 stars0 ratingsThe Right’s First Amendment: The Politics of Free Speech & the Return of Conservative Libertarianism Rating: 0 out of 5 stars0 ratingsThe Prisonhouse of Nations - America! Rating: 0 out of 5 stars0 ratingsSummary of Wendy Hui Kyong Chun's Discriminating Data Rating: 0 out of 5 stars0 ratingsMaking Human Dignity Central to International Human Rights Law: A Critical Legal Argument Rating: 0 out of 5 stars0 ratingsFrom Parchment to Dust: The Case for Constitutional Skepticism Rating: 0 out of 5 stars0 ratingsComments on Thomas Hobbes Book (1651) The Leviathan Part 1 Rating: 0 out of 5 stars0 ratingsConstructing Basic Liberties: A Defense of Substantive Due Process Rating: 3 out of 5 stars3/5Virtual Freedom: Net Neutrality and Free Speech in the Internet Age Rating: 4 out of 5 stars4/5Courts Liberalism And Rights: Gay Law And Politics In The United States and Canada Rating: 0 out of 5 stars0 ratingsThe Bricks before Brown: The Chinese American, Native American, and Mexican Americans' Struggle for Educational Equality Rating: 0 out of 5 stars0 ratingsAlternative to Darwinism and Creationism Based on Free Will Rating: 0 out of 5 stars0 ratingsOn Digital Advocacy: Saving the Planet While Preserving Our Humanity (Speaker's Corner) Rating: 0 out of 5 stars0 ratingsFree Speech On Trial: Communication Perspectives on Landmark Supreme Court Decisions Rating: 0 out of 5 stars0 ratingsEternally Vigilant: Free Speech in the Modern Era Rating: 0 out of 5 stars0 ratingsEthics (Serapis Classics) Rating: 0 out of 5 stars0 ratingsData Personified: How Fraud Is Transforming the Meaning of Identity Rating: 0 out of 5 stars0 ratingsImmigration Matters: Movements, Visions, and Strategies for a Progressive Future Rating: 0 out of 5 stars0 ratingsJurisprudence (Essay) Rating: 0 out of 5 stars0 ratingsDifference without Domination: Pursuing Justice in Diverse Democracies Rating: 0 out of 5 stars0 ratingsThe Myth of the Closed Mind: Understanding Why and How People Are Rational Rating: 0 out of 5 stars0 ratingsData Cartels: The Companies That Control and Monopolize Our Information Rating: 0 out of 5 stars0 ratingsSecular Government, Religious People Rating: 0 out of 5 stars0 ratingsSustainability Management: Lessons from and for New York City, America, and the Planet Rating: 5 out of 5 stars5/5Social Policy in the United States: Future Possibilities in Historical Perspective Rating: 4 out of 5 stars4/5The Constitutional Convention of 1787: Constructing the American Republic Rating: 0 out of 5 stars0 ratingsBetween Peace and War: 40th Anniversary Revised Edition Rating: 0 out of 5 stars0 ratingsWhy?: The Philosophy Behind the Question Rating: 0 out of 5 stars0 ratings
Technology & Engineering For You
Elon Musk: Tesla, SpaceX, and the Quest for a Fantastic Future Rating: 4 out of 5 stars4/5The Art of War Rating: 4 out of 5 stars4/580/20 Principle: The Secret to Working Less and Making More Rating: 5 out of 5 stars5/5Electrical Engineering 101: Everything You Should Have Learned in School...but Probably Didn't Rating: 5 out of 5 stars5/5Understanding Media: The Extensions of Man Rating: 4 out of 5 stars4/5Ultralearning: Master Hard Skills, Outsmart the Competition, and Accelerate Your Career Rating: 4 out of 5 stars4/5The ChatGPT Millionaire Handbook: Make Money Online With the Power of AI Technology Rating: 0 out of 5 stars0 ratingsThe Big Book of Maker Skills: Tools & Techniques for Building Great Tech Projects Rating: 4 out of 5 stars4/5The Big Book of Hacks: 264 Amazing DIY Tech Projects Rating: 4 out of 5 stars4/5How to Write Effective Emails at Work Rating: 4 out of 5 stars4/5The CIA Lockpicking Manual Rating: 5 out of 5 stars5/5The 48 Laws of Power in Practice: The 3 Most Powerful Laws & The 4 Indispensable Power Principles Rating: 5 out of 5 stars5/5The Art of War Rating: 4 out of 5 stars4/5Pilot's Handbook of Aeronautical Knowledge (Federal Aviation Administration) Rating: 4 out of 5 stars4/5The Systems Thinker: Essential Thinking Skills For Solving Problems, Managing Chaos, Rating: 4 out of 5 stars4/5U.S. Marine Close Combat Fighting Handbook Rating: 4 out of 5 stars4/5My Inventions: The Autobiography of Nikola Tesla Rating: 4 out of 5 stars4/5Broken Money: Why Our Financial System is Failing Us and How We Can Make it Better Rating: 5 out of 5 stars5/5Smart Phone Dumb Phone: Free Yourself from Digital Addiction Rating: 0 out of 5 stars0 ratingsThe Fast Track to Your Technician Class Ham Radio License: For Exams July 1, 2022 - June 30, 2026 Rating: 5 out of 5 stars5/5Logic Pro X For Dummies Rating: 0 out of 5 stars0 ratingsHow to Disappear and Live Off the Grid: A CIA Insider's Guide Rating: 0 out of 5 stars0 ratingsRust: The Longest War Rating: 4 out of 5 stars4/5No Nonsense Technician Class License Study Guide: for Tests Given Between July 2018 and June 2022 Rating: 5 out of 5 stars5/5The Complete Titanic Chronicles: A Night to Remember and The Night Lives On Rating: 4 out of 5 stars4/5A Night to Remember: The Sinking of the Titanic Rating: 4 out of 5 stars4/5
Reviews for Machine See, Machine Do
0 ratings0 reviews
Book preview
Machine See, Machine Do - Patrick K. Lin
Contents
Introduction
Ghost in the Machine
Franglen’s Monster
Automating Bias
Automating the Watchers
Lights, Camera, Surveillance!
Cloudy Crystal Ball: Predictive Policing
Smile! You’re on Camera: Video Surveillance
AI Spy with My Little Eye: Facial Recognition
False Positive
Smoking Gun or Smoke and Mirrors: Forensic Evidence
Beyond a Relative Doubt: Forensic DNA Databases & Familial DNA Searches
May It Please the Algorithm
Algorithms Get Their Day in Court
Behind Closed Source: How the Law Keeps Algorithms Hidden from Us
Digital Shackles: How Parole Apps Put Prison in Your Pocket
Troubleshooting the Problem
Futureproofing Our Civil Liberties
The Wild, Wild West: The Role of Government
Flipping the Script: The Role of the Private Sector
Beyond Technology: Expecting More from AI
Acknowledgments
Appendix
For every person who has ever been shortchanged, excluded, and underestimated by machines.
For the public interest technologists improving the world with their hope, vision, and perseverance.
For the activists and advocates protecting civil liberties through organizing, education, and dedication.
Introduction
Much of what New York City looks like today is attributed to a man who never held elected office or received any formal training in architecture or urban planning.
Robert Moses has been called the master builder
of mid-twentieth century New York and its surrounding suburbs. He shaped much of the infrastructure of modern New York City, Long Island, Rockland County, and Westchester County (Caro, 1974).
Over the course of his forty-four-year career, Moses built nearly seven hundred miles of road, including massive highways that stretched out of Manhattan into Long Island and Upstate New York; twenty thousand acres of parkland and public beaches; 658 playgrounds; seven new bridges; the UN Headquarters; the Central Park Zoo; and the Lincoln Center for the Performing Arts (Burkeman, 2015). It would be an understatement to say Moses left a lasting mark on New York. In the twentieth century, the influence of Robert Moses on the cities of America was greater than that of any other person,
wrote American historian Lewis Mumford.
However, new, large-scale developments come with a price—and not everyone pays the same amount.
To build hundreds of miles of highways and dozens of housing and urban renewal projects, Moses had more than five hundred thousand people evicted (Gratz, 2007). Black and Brown people comprised 40 percent of the evicted population at a time when those demographics made up only about 10 percent of the New York City’s overall population (Census Bureau, 2021). The construction of Lincoln Center alone displaced more than seven thousand working-class families and eight hundred businesses. Many of these evicted New Yorkers ended up in Harlem and the Bronx, further segregating the city (Williams, 2017). Moses also avoided building public pools in Black neighborhoods and instead designed those same neighborhoods to be prone to traffic congestion, not only withholding public goods from Black neighborhoods, but also forcing them to bear the brunt of the social costs (Schindler, 2015).
Robert Moses with a model of the proposed Battery Bridge. Source: The Library of Congress.
Moses infamously hated the idea of poor people—particularly poor people of color—using the new public parks and beaches he was building on Long Island (Burkeman, 2015). To that end, Moses used his influence and connections to pass a law forbidding public buses on highways, but he knew laws could someday be repealed. Legislation can always be changed,
Moses said. It’s very hard to tear down a bridge once it’s up.
So Moses built scores of bridges that were too low to let public buses pass, literally concretizing discrimination
(Bornstein, 2017). The effect of these decisions has been profound and enduring. Decades later, the bus laws Moses fought for were overturned. Still, the towns he built along the highways remain as segregated as ever.
People often do not want to believe seemingly innocuous objects—like bridges or highways—can be racial or political, but as Moses’ buildings and plans show, human history is inherently racial and political. Moses’ racist views played out in what he built, how he built, and where he built.
But Moses was not alone. He wielded tremendous power and influence throughout his career, but he was still just an individual operating within a system built on bias and racism. For example, the Federal Housing Administration’s Underwriting Manual states incompatible racial groups should not be permitted to live in the same communities,
recommending highways be built as a way to separate Black neighborhoods from white neighborhoods (Gross, 2017). Rooting out bias isn’t only about powerful individuals; it isn’t even just about you or me. It’s about history and systems that continue to exist, bridges that are too difficult to tear down.
Discriminatory decisions and policies of the past impact the present. Racial and social inequity affect the very fabric of our reality. Everything has costs and benefits, and these are not evenly distributed. The decision, whether conscious or unconscious, to advance or burden some members of society over others is fundamentally racial and political.
Artificial intelligence is no different. The technology is relentlessly improving and increasingly pervasive, yet despite well-documented biases, AI developed in the private and public sectors alike consistently fail to account for it. Somehow, in the past two decades, we got the idea machines make better decisions than humans. We began saying things like, People are biased, but AI will be more objective.
We have forgotten humans design and deploy AI to serve their purposes. Humans, even those with the best intentions, can introduce bias to the AI they develop. Technology is not inherently objective or fair.
Today’s technology, built from yesterday’s data, will reflect the biased environment from which that data came. Bias often appears in AI systems through factors like race and gender, which generally are not directly inputted into the system but still have a strong influence over the system’s decisions. The system is especially prone to bias when one of these factors is strongly correlated with information directly used by the system.
For example, suppose a system that makes determinations about someone’s level of education uses zip code as a factor to make its decisions. Direct information about race is never given to the system, so how can a system like that be biased?
Zip code is correlated with race since a lot of neighborhoods in America are still segregated,
senior staff technologist Daniel Kahn Gillmor at the ACLU’s Speech, Privacy, and Technology Project said to me. Gillmor’s work focuses on the way our technical infrastructure shapes society and impacts civil liberties. The data you’re using to make these guesses is ultimately going to be pulled from a society that has a bunch of other problems, and the system is going to just reflect those problems back.
By using zip code as a factor, the AI system is indirectly making decisions based on race. In other words, zip code is a proxy for race. Therefore, even if the system’s math and logic is all correct, an underlying ethical question reveals itself: is it appropriate to make these decisions based on these inputs?
A magical machine offering the promise of objectivity and fairness is extremely appealing. The public can be tricked into accepting an imperfect or even incompetent algorithm, particularly when the current state of an institution has historically been plagued by prejudice and bias, like the judicial system. We know things need to change and we want to believe technology can be that change. However, unlike humans, an algorithm cannot dissent, disobey, or make exceptions. People, on the other hand, can learn to account for ways in which data is a representation of the past.
So a question at the center of algorithmic fairness is whether an algorithm can be designed to comprehend the social and historical context of the data it relies on. Machines that cannot understand context will merely pass on institutionalized biases. To borrow a computer science adage: bias in, bias out.
Algorithms pervade our lives today. However, the development and deployment of AI is virtually unregulated. Nowhere is this lack of regulation more problematic than in the criminal justice system, where AI directs the policing of our streets, surveils our citizens, and determines whether people should go to jail.
From my independent research and experiences at organizations like the Legal Aid Society’s DNA Unit and the ACLU’s Speech, Privacy, and Technology Project, I have seen time and time again how our institutions have relied more and more on automating justice to the detriment of our civil liberties.
Over time, something became clear to me: AI isn’t going to take over the world, at least not in the Terminator-style apocalypse we might think. Instead, if we are not careful, AI will take all our human mistakes and immortalize them. Fixing the technology is just one part of the solution. Ultimately, we need to fix our systems and institutions. We need to think critically about the way humans use AI on other humans.
Despite the hallowed way people have used words like algorithm
and AI,
almost every aspect of such decision-making is ultimately left to humans. Algorithms are designed by humans and run on computers built by humans. Algorithms are trained on data created and collected by humans. Algorithms are evaluated based on how well they reflect human priorities and values.
Plenty of literature on AI focuses on how the technology is flawed and how algorithmic bias must be addressed. I want to focus on the human aspect of this technology: the people who design and deploy AI and the history of how it came to be. I believe fixing bias in AI is about changing the way humans treat other humans.
My goal is to provide a new perspective on how to tackle this fascinating, complex, and important problem which exists at the intersection of the future of technology and our civil liberties. For those just starting to learn about AI and algorithmic bias, I hope this book can be an approachable guide to this space.
We need to think critically about how some of the same technology that has made our lives more convenient has also been used in unexpected, invasive, and cruel ways, especially on human lives who are historically neglected, marginalized, and victimized. We cannot expect simple and elegant solutions to these messy and complicated problems. Machines inherit our views and our history, including our prejudices and biases.
For policymakers, I hope this book shows just how lawless the realm of AI is, particularly as it pertains to criminal justice. It is no secret law and policy have failed to keep pace with technological advances. However, losing this race will diminish our individual rights.
This technology is making decisions that have profound ramifications, placing more people in the crosshairs of the police, and mislabeling people as criminals for factors they have no control over. I also highlight potential solutions and provide a more complete understanding of how law and policy can begin to right some of these wrongs and prevent future harm.
For AI developers and software engineers, I hope this book emphasizes how crucial your role is in both the problem and the solution to algorithmic bias. As AI evolves, your role shifts further from developer or engineer—and closer to policymaker. The lines of code you are writing and the products you are creating affect real-life people. It is important to recognize how that impact will vary for each person. Law and policy are just one piece of the puzzle. The way we build our technology must also change.
Ultimately, we need to ensure these processes do not lose their humanity. Our government’s once individualized forms of surveillance have become mass surveillance. Our police and courts are turning to machines to make significant, life-altering decisions. We’ve long given up the idea justice is blind. It’s about time we give up the idea technology is blind too.
AI is not some kind of silver bullet. We cannot rely solely on machines to solve problems created by humans. When autopilot was developed, we did not send passengers on airplanes without pilots in the cockpit. Similarly, we cannot completely remove human perspective or interaction from processes like policing and criminal justice, processes that have immense human impact. Machines are not a substitute for community engagement and holistic crime reduction measures.
Ironically (and tragically), the human obsession with predicting the future results in technology recreating the past—and its mistakes. The past cannot be rewritten, but one way or another, the response to AI in surveillance and criminal justice will determine whether hard-won civil liberties endure or become forgotten relics.
Part 1:
Ghost in the Machine
Chapter 1
Franglen’s Monster
Pray, Mr. Babbage, if you put into the machine wrong figures, will the right answers come out?
— Charles Babbage, designer of the Analytical Engine, a prototype of the first computer
I’ve loved science fiction movies since I first watched C-3PO awkwardly shuffle through the sterile, white hallways of a spaceship in Star Wars: Episode IV – A New Hope (1977). I was especially fascinated by movie depictions of AI and its relationship with human characters. AI and its advancements may have become a popular topic in recent years, but AI has been a focal point for filmmakers for nearly a century (Tomlinson, 2018).
AI’s first onscreen appearance was in the 1927 German expressionist movie Metropolis, in which a humanoid robot wreaks havoc in the titular city. Not a great first impression for AI, I’ll admit. Although the first AI to appear in an American movie wouldn’t be for another twenty-four years, the portrayal was noticeably more positive. Gort from the 1951 movie The Day the Earth Stood Still was a silent guardian to the movie’s protagonist.
In 1968, HAL 9000 from 2001: A Space Odyssey was the main antagonist but was far more human than any other movie robot before him, despite being a fixture with no body. In 1983’s WarGames, it was the first depiction of AI’s involvement in a nonfictional war. A year later, The Terminator gave us Skynet,
and to this day, Skynet
is a commonly used analogy for the threat posed by advanced AI.
In 1999, The Matrix graced