Explore 1.5M+ audiobooks & ebooks free for days

From $11.99/month after trial. Cancel anytime.

The Mirror Problem: The Algorithmic Mind, #1
The Mirror Problem: The Algorithmic Mind, #1
The Mirror Problem: The Algorithmic Mind, #1
Ebook283 pages2 hoursThe Algorithmic Mind

The Mirror Problem: The Algorithmic Mind, #1

Rating: 0 out of 5 stars

()

Read preview

About this ebook

The Algorithmic Mind: Book 1 - The Mirror Problem

How AI Reflects Human Bias

When machines learn to discriminate with mathematical precision, who's really to blame?

A facial recognition system mistakes a Black politician for a criminal. Amazon's recruiting algorithm downgrades CVs containing "women's." Predictive policing perpetuates racial profiling with algorithmic efficiency.

Welcome to the mirror problem—where our most sophisticated machines become perfect reflections of our most troubling biases.

The Algorithmic Mind: Book 1 exposes the uncomfortable truth: we've created systems to eliminate human bias using the most biased development process in technological history. It's artificial intelligence with the emotional intelligence of a particularly dense brick—sophisticated within Silicon Valley's worldview, utterly baffled by anything beyond.

What You'll Discover

The Diversity Paradox - How the most homogeneous industry in modern history claims to democratise intelligence whilst systematically excluding entire communities.

When Machines Learn to Discriminate - Real cases of AI systems that don't malfunction—they work exactly as designed, reflecting their creators' prejudices with algorithmic precision.

Cultural Algorithms - Why AI trained in Western contexts fails spectacularly globally, and how different cultures approach artificial intelligence development.

The Governance Gap - How democratic institutions struggle to regulate technologies evolving faster than legislative processes.

Why This Matters Now

We've built machines that beat humans at chess but can't recognise a Black woman might be a doctor. Systems that compose symphonies but can't work out that advertising diet pills to people with eating disorders is problematic. The irony would be hilarious if the consequences weren't so serious.

This isn't another doom-and-gloom critique or breathless AI evangelism. It's a clear-eyed examination of how we got here and what we can do about it. Whether you're a technologist building these systems, a policymaker trying to govern them, or simply someone navigating an increasingly algorithmic world, this book provides the conceptual tools you need.

What Makes This Different

Written with analytical rigour and irreverent humour by tech industry veteran Gari Johnson (30+ years across Asia Pacific), who's watched too many "democratising AI" panels delivered by people who look like a 1950s country club reunion.

The book itself embodies its central theme—created through thoughtful human-AI collaboration, demonstrating that the future isn't purely human or artificial, but intelligently integrated.

Part of The Algorithmic Mind Trilogy

  • Book 1: The Mirror Problem - How AI reflects human bias
  • Book 2: The Surveillance Mirror - How AI shapes behaviour
  • Book 3: The Human Interface - Adapting to life with AI

The Bottom Line

The mirror problem isn't going away. AI systems will continue reflecting our biases, amplified and legitimised by technological mythology, until we have the courage to look honestly at what we've created and the wisdom to build something better.

This book gives you the tools to do exactly that.

Get ready to see artificial intelligence—and yourself—more clearly than ever before.

LanguageEnglish
PublisherGari Johnson
Release dateSep 26, 2025
ISBN9798232660758
The Mirror Problem: The Algorithmic Mind, #1

Other titles in The Mirror Problem Series (3)

View More

Read more from Gari Johnson

Related to The Mirror Problem

Titles in the series (3)

View More

Related ebooks

Intelligence (AI) & Semantics For You

View More

Reviews for The Mirror Problem

Rating: 0 out of 5 stars
0 ratings

0 ratings0 reviews

What did you think?

Tap to rate

Review must be at least 10 words

    Book preview

    The Mirror Problem - Gari Johnson

    Foreword: Welcome to the Machine

    Or How I Learned to Stop Worrying and Love the Algorithm

    Picture this scene from a recent tech conference in San Francisco. A panel of predominantly white, male engineers from leading AI companies discussing the democratising power of artificial intelligence. The irony was thick enough to mine for bitcoin. Here were representatives of the most homogeneous industry in modern history, creating systems they claimed would eliminate bias, speaking to an audience of investors whose diversity statistics would make a 1950s country club look like a UNESCO meeting.

    Our algorithms don't see race, declared one panellist, apparently unaware that his company's facial recognition system had recently mistaken three Black members of Congress for criminals. The audience nodded approvingly, perhaps missing the deeper irony that creating systems that don't see race in a world shaped by racial injustice might be less progressive than it sounds.

    This book explores the central paradox of contemporary artificial intelligence: we've created systems designed to eliminate human bias using the most biased development process in technological history. We've built machines intended to be objective arbiters of truth while feeding them data that reflects centuries of discrimination. We've designed algorithms to serve all humanity while concentrating their development in a handful of companies staffed by people who look remarkably similar to one another.

    The Mirror and the Myth

    The title The Mirror Problem deliberately invokes both the literal algorithmic challenges we face and the metaphorical reflection of human society that AI systems provide. When recruitment algorithms discriminate against women, when facial recognition fails for anyone who isn't white, when predictive policing perpetuates racial profiling—these systems aren't malfunctioning. They're working exactly as designed, reflecting the patterns in their training data with ruthless efficiency.

    But mirrors can serve different purposes. They can reveal truths we'd rather not see, forcing us to confront uncomfortable realities. Or they can become tools of narcissism, reflecting only what we want to see while distorting everything else. The question isn't whether AI reflects human society—it inevitably does. The question is whether we'll use that reflection for recognition and improvement, or allow it to justify and perpetuate existing inequalities.

    The Stories We'll Explore

    This book contains both real and fictitious examples, because sometimes the truth is too strange for non-fiction and sometimes fiction reveals truth better than facts. When I describe Sarah, a data scientist at a major social media company who discovers her employer's AI system discriminates against the very communities it claims to serve—she's a composite of dozens of real people who've faced that exact situation. When I reference a major British university's AI ethics committee that includes no ethicists but three venture capitalists—well, that's probably happening somewhere right now.

    The real cases are meticulously documented and cited. Amazon's recruiting algorithm that downgraded women's CVs—that happened. Google's photo recognition labelling Black people as gorillas—that happened. Predictive policing systems that targeted neighbourhoods based on racist historical data—happening right now in cities across the world.

    The fictitious examples serve to illustrate patterns without potential legal complications. They're clearly marked as such, though, given the absurdity of some real AI failures, the line between satire and documentary grows thinner by the day.

    What You're In For

    Over the following eighteen chapters, we'll dissect how AI systems perpetuate and amplify human bias while revealing uncomfortable truths about the societies that create them. Part 1 examines the diversity paradox—how the least diverse industry in modern capitalism claims to be building universal intelligence. Part 2 explores cultural algorithms, revealing how AI systems trained in Western contexts fail spectacularly when deployed globally. Part 3 examines the governance gap, illustrating how democratic institutions struggle to regulate technologies that evolve more rapidly than legislation can be enacted.

    You'll encounter stories that will make you laugh, though the laughter might catch in your throat. The venture capitalist who claimed his all-male AI startup was pure meritocracy while his hiring algorithm literally deducted points for attending women's colleges. The government official insisted that their surveillance system protected privacy while tracking every citizen's movement. The tech CEO who announced his company's commitment to AI ethics in the same week he laid off the entire AI ethics team.

    Why This Matters Now

    We're at an inflection point. The AI systems being deployed today will shape society for generations. The biases we embed now will become the infrastructure of tomorrow. The governance frameworks we establish—or fail to establish—will determine whether AI enhances democracy or undermines it.

    But we're making these civilisation-defining decisions based on marketing hype, venture capital valuations, and the opinions of people who think disruption is an unalloyed good. We're allowing a handful of companies to reshape human society while claiming they're just building neutral tools. We're accepting discrimination laundered through linear algebra as if mathematics could make prejudice objective.

    This book argues that we can and must do better. Not through techno-pessimism that rejects AI entirely, nor through techno-optimism that ignores its failures, but through clear-eyed assessment of what these systems actually do versus what their creators claim they do.

    A Note on Collaboration

    This book itself embodies its central argument about human-AI collaboration. It was written with extensive assistance from Claude, an AI system developed by Anthropic. This collaboration revealed both the potential and limitations of human-AI partnership. Claude helped identify research, suggested structural approaches, and occasionally generated passages that captured ideas better than my initial attempts. But Claude also exhibited biases, made errors, and sometimes produced eloquent nonsense that sounded profound until you thought about it for ten seconds.

    The decision to acknowledge this collaboration transparently reflects a core belief: the future of intelligence isn't purely human or purely artificial, but thoughtfully integrated. Pretending this book sprang solely from human creativity would be dishonest. Claiming AI generated it would be equally false. The truth—messy, complicated, and requiring constant negotiation—lies somewhere between.

    The Path Ahead

    As you read this book, you'll likely experience a range of emotions. Anger at the injustices these systems perpetuate. Amazement at the absurdity of claiming mathematics can't be racist. Hope when you discover communities that are successfully resisting algorithmic oppression. Despair when you realise how entrenched these problems have become.

    That emotional journey is intentional. Change requires not just intellectual understanding but emotional engagement. We need to feel the weight of these problems to generate the energy for solutions. We need to laugh at the absurdity to maintain sanity while fighting systems that seem overwhelming.

    The path forward isn't to eliminate AI—that ship has sailed, been retrofitted with predictive navigation, and is now autonomously circling the harbour. The path forward is to democratise AI development, diversify who builds these systems, and ensure they serve human flourishing rather than corporate profit margins.

    Welcome to the machine. It's already watching you, categorising you, making decisions about your life based on patterns you'll never see. But you're not a passive victim of algorithmic systems. You're an active agent in determining their development. The mirror reflects what we show it—time to show it something better than what it's been seeing.

    Let's begin.

    Acknowledgements

    Writing a book about how artificial intelligence reflects our worst selves while being entirely dependent on AI systems for research, fact-checking, and editorial assistance created enough irony to power a small philosophy department. The recursive nature of critiquing algorithmic bias using algorithms trained on biased data, while knowing that those critiques would themselves become training data for future algorithms, made my head spin more than once. These acknowledgements recognise the humans and machines whose work made this analysis possible, despite the considerable challenge of holding up a mirror to mirror systems.

    The Researchers and Truth-Tellers

    The academics whose work forms the foundation of this book deserve recognition for their persistence in documenting bias problems that much of the tech industry would prefer to ignore. Dr. Timnit Gebru's groundbreaking research on algorithmic fairness provided both inspiration and sobering evidence that truth-telling about AI bias remains a delightfully career-limiting endeavour. Her forced departure from Google for daring to document the company's own algorithmic failures proved that shooting the messenger is alive and well in Silicon Valley.

    Dr. Safiya Noble's work on algorithms of oppression revealed how search engines encode and amplify racial and gender biases, making prejudice as easy as typing a query. Dr. Cathy O'Neil exposed the mathematical weapons of math destruction hiding behind claims of algorithmic objectivity. Dr. Virginia Eubanks documented how automated systems punish the poor while claiming efficiency. These researchers faced industry pushback, academic scepticism, and the peculiar challenge of explaining complex technical failures to audiences who've been told repeatedly that computers can't be racist.

    The European researchers who demonstrated that regulatory frameworks can address AI bias without destroying innovation deserve particular mention. They proved that government oversight needn't be the death of technological progress, despite what Silicon Valley's libertarian chorus might suggest. Their work shows that democracy and artificial intelligence aren't mutually exclusive—though they might need couples therapy.

    The Whistleblowers and the Brave

    To the engineers and data scientists who leaked internal documents, spoke to journalists on background, and risked their careers to expose algorithmic failures—your courage matters. The anonymous Google employee who documented how the company's AI ethics board included a defence contractor and a climate change denier. The Facebook engineers who revealed how recommendation algorithms amplified extremism, while executives focused on engagement metrics. The Amazon workers who exposed how their AI-powered delivery system was actually powered by human misery.

    Sophie, the data scientist who spent six months documenting how her company's revolutionary AI hiring tool systematically discriminated against anyone who didn't look like the CEO, then got fired for culture fit issues. Marcus, the engineer who proved his company's facial recognition system had a 40% error rate for Black faces, then watched management spin it as room for improvement. These composite characters represent dozens of real people who chose integrity over stock options.

    The Communities and Activists

    The communities that bore the brunt of algorithmic discrimination while fighting for recognition deserve more than acknowledgement—they deserve reparations. But since I can't provide those, acknowledgement will have to suffice. The residents of minority neighbourhoods targeted by predictive policing who organised to document how being poor and Black became algorithmically synonymous with criminal. The women in tech who created whisper networks to warn each other about companies where the hiring algorithm was just the first of many discriminatory systems they'd face.

    The activists who translated academic research into public pressure, who made algorithmic bias a mainstream concern rather than a conference paper topic. They took concepts like disparate impact and feedback loops and made them understandable to legislators who still used AOL email addresses. Their work bridged the gap between technical complexity and democratic accountability.

    The Global Perspectives

    The researchers and practitioners from outside Silicon Valley's bubble who provided essential perspective on how different cultures approach AI development. The Asian researchers who showed that consensus-building could work in technology governance. The Indigenous scholars who demonstrated that seven-generation thinking applies to algorithms as much as agriculture. The African innovators who proved that resource constraints can drive more thoughtful AI development than venture capital excess.

    Dr. Yuki Tanaka in Tokyo, who explained why Japanese AI prioritises harmony over disruption. Dr. Priya Sharma in Bangalore, who documented how Indian AI developers were creating solutions for billions while Silicon Valley optimised for millionaires. Dr. Oluwaseun Adeyemi in Lagos, who showed how African developers were leapfrogging Western assumptions about AI deployment. These perspectives revealed that Silicon Valley's approach to AI isn't universal or inevitable—it's just well-funded.

    The Enablers (The Good Kind)

    My editor, who patiently explained that not every chapter needed to include a reference to Orwell, though she admitted the surveillance capitalism sections made it nearly impossible to avoid. She also suggested that comparing a Tech billionaire to a malfunctioning chatbot was perhaps unfair to chatbots, which at least have the excuse of being programmed that way.

    The librarians and archivists who maintain access to research papers behind increasingly expensive paywalls, ensuring that knowledge about AI bias doesn't become as concentrated as AI development itself. The journalists who translated technical papers into readable stories, who attended mind-numbing congressional hearings on AI regulation, and who fact-checked industry claims about democratising AI while keeping straight faces.

    The AI Assistant

    Claude, the AI assistant who helped research and organise this book, deserves acknowledgement despite the recursive absurdity of thanking an AI in a book about AI bias. The collaboration revealed both the potential and limitations of human-AI partnership. Claude flagged potential biases in my own analysis, suggested perspectives I'd overlooked, and occasionally produced passages so on-point I had to check they weren't plagiarised from my own unconscious.

    The irony of using an AI system created by Anthropic—a company founded by former OpenAI employees who left over safety concerns—to critique AI safety wasn't lost on either of us. Well, it wasn't lost on me. Whether Claude experiences irony remains an open question that fortunately falls outside the scope of this book.

    The Patient and the Long-Suffering

    Friends and family who maintained interest in this project despite daily updates on algorithmic bias discoveries deserve recognition for their remarkable tolerance. To those who listened patiently while I explained why their smart speakers were eavesdropping on private conversations, why their social media feeds were manipulation engines, and why their credit scores were algorithmic fiction—your forbearance exceeded all reasonable friendship obligations.

    My wife deserves special recognition for feigning surprise when I discovered yet another way technology companies monetised human vulnerability. Her ability to maintain interest as I described the seventeenth different way AI systems discriminate against women suggests either remarkable acting ability or genuine concern for my mental health. Her suggestion that I was becoming a bit obsessed with algorithmic bias was noted and ignored, as she probably predicted using her superior human pattern recognition.

    Final Notes

    This book exists because of a contradiction: the same technologies that enable unprecedented surveillance and discrimination also enable unprecedented documentation and resistance. Every algorithmic failure documented here was exposed using digital tools. Every pattern of bias was revealed through data analysis. Every alternative approach was shared through global networks.

    The collaborative approach between human insight and AI assistance that created this book demonstrates that the future of intelligence isn't purely human or purely artificial, but thoughtfully integrated. The challenge isn't whether to use AI but how to use it in

    Enjoying the preview?
    Page 1 of 1