Just last week, Google was forced to pump the brakes on its AI image generator, called Gemini, after critics complained that it was pushing bias ... against white people.

The controversy started with — you guessed it — a viral post on X. According to that post from the user @EndWokeness, when asked for an image of a Founding Father of America, Gemini showed a Black man, a Native American man, an Asian man, and a relatively dark-skinned man. Asked for a portrait of a pope, it showed a Black man and a woman of color. Nazis, too, were reportedly portrayed as racially diverse.

After complaints from the likes of Elon Musk, who called Gemini’s output “racist” and Google “woke,” the company suspended the AI tool’s ability to generate pictures of people.

“It’s clear that this feature missed the mark. Some of the images generated are inaccurate or even offensive,” Google Senior Vice President Prabhakar Raghavan wrote, adding that Gemini does sometimes “overcompensate” in its quest to show diversity.

Raghavan gave a technical explanation for why the tool overcompensates: Google had taught Gemini to avoid falling into some of AI’s classic traps, like stereotypically portraying all lawyers as men. But, Raghavan wrote, “our tuning to ensure that Gemini showed a range of people failed to account for cases that should clearly not show a range.”

This might all sound like just the latest iteration of the dreary culture war over “wokeness” — and one that, at least this time, can be solved by quickly patching a technical problem. (Google plans to relaunch the tool in a few weeks.)

But there’s something deeper going on here. The problem with Gemini is not just a technical problem.

It’s a philosophical problem — one for which the AI world has no clear-cut solution.

What does bias mean?

Imagine that you work at Google. Your boss tells you to design an AI image generator. That’s a piece of cake for you — you’re a brilliant computer scientist! But one day, as you’re testing the tool, you realize you’ve got a conundrum.

You ask the AI to generate an image of a CEO. Lo and behold, it’s a man. On the one hand, you live in a world where the vast majority of CEOs are male, so maybe your tool should accurately reflect that, creating images of man after man after man. On the other hand, that may reinforce gender stereotypes that keep women out of the C-suite. And there’s nothing in the definition of “CEO” that specifies a gender. So should you instead make a tool that shows a balanced mix, even if it’s not a mix that reflects today’s reality?

This comes down to how you understand bias.

Computer scientists are used to thinking about “bias” in terms of its statistical meaning: A program for making predictions is biased if it’s consistently wrong in one direction or another. (For example, if a weather app always overestimates the probability of rain, its predictions are statistically biased.) That’s very clear, but it’s also very different from the way most people use the word “bias” — which is more like “prejudiced against a certain group.”

The problem is, if you design your image generator to make statistically unbiased predictions about the gender breakdown among CEOs, then it will be biased in the second sense of the word. And if you design it not to have its predictions correlate with gender, it will be biased in the statistical sense.

So how should you resolve the trade-off?

“I don’t think there can be a clear answer to these questions,” Julia Stoyanovich, director of the NYU Center for Responsible AI, told me when I previously reported on this topic. “Because this is all based on values.”

Embedded within any algorithm is a value judgment about what to prioritize, including when it comes to these competing notions of bias. So companies have to decide whether they want to be accurate in portraying what society currently looks like, or promote a vision of what they think society could or even should look like — a dream world.

How can tech companies do a better job navigating this tension?

The first thing we should expect companies to do is get explicit about what an algorithm is optimizing for: Which type of bias will it focus on reducing? Then companies have to figure out how to build that into the algorithm.

Part of that is predicting how people are likely to use an AI tool. They might try to create historical depictions of the world (think: white popes) but they might also try to create depictions of a dream world (female popes, bring it on!).

“In Gemini, they erred towards the ‘dream world’ approach, understanding that defaulting to the historic biases that the model learned would (minimally) result in massive public pushback,” wrote Margaret Mitchell, chief ethics scientist at the AI startup Hugging Face.

Google might have used certain tricks “under the hood” to push Gemini to produce dream-world images, Mitchell explained. For example, it may have been appending diversity terms to users’ prompts, turning “a pope” into “a pope who is female” or “a Founding Father” into “a Founding Father who is Black.”

But instead of adopting only a dream-world approach, Google could have equipped Gemini to suss out which approach the user actually wants (say, by soliciting feedback about the user’s preferences) — and then generate that, assuming the user isn’t asking for something off-limits.

What counts as off-limits comes down, once again, to values. Every company needs to explicitly define its values and then equip its AI tool to refuse requests that violate them. Otherwise, we end up with things like Taylor Swift porn.

AI developers have the technical ability to do this. The question is whether they’ve got the philosophical ability to reckon with the value choices they’re making — and the integrity to be transparent about them.

This story appeared originally in Today, Explained, Vox’s flagship daily newsletter. Sign up here for future editions.

$5/month

Yes, I'll give $5/month

Yes, I'll give $5/month

We accept credit card, Apple Pay, and Google Pay. You can also contribute via