Google has admitted that its Gemini AI model “missed the mark” after a flurry of criticism about what many perceived as “anti-white bias.” Numerous users reported that the system was producing images of people of diverse ethnicities and genders even when it was historically inaccurate to do so. The company said Thursday it would “pause” the ability to generate images of people until it could roll out a fix.
When prompted to create an image of Vikings, Gemini showed exclusively Black people in traditional Viking garb. A “founding fathers” request returned Indigenous people in colonial outfits; another result depicted George Washington as Black. When asked to produce an image of a pope, the system showed only people of ethnicities other than white. In some cases, Gemini said it could not produce any image at all of historical figures like Abraham Lincoln, Julius Caesar, and Galileo.
Many right-wing commentators have jumped on the issue to suggest this is further evidence of an anti-white bias among Big Tech, with entrepreneur Mike Solana writing that “Google’s AI is an anti-white lunatic.”
But the situation mostly highlights that generative AI systems are just not very smart.
“I think it is just lousy software,” Gary Marcus, an emeritus professor of psychology and neural science at New York University and an AI entrepreneur, wrote on Wednesday on Substack.
Google launched its Gemini AI model two months ago as a rival to the dominant GPT model from OpenAI, which powers ChatGPT. Last week Google rolled out a major update to it with the limited release of Gemini Pro 1.5, which allowed users to handle vast amounts of audio, text, and video input.
Gemini also created images that were historically wrong, such as one depicting the Apollo 11 crew that featured a woman and a Black man.
On Wednesday, Google admitted its system was not working properly.
“We’re working to improve these kinds of depictions immediately,” Jack Krawczyk, a senior director of product management at Google’s Gemini Experiences, told WIRED in an emailed statement. “Gemini’s AI image generation does generate a wide range of people. And that’s generally a good thing because people around the world use it. But it’s missing the mark here.”
Krawczyk explained the situation further in a post on X: “We design our image generation capabilities to reflect our global user base, and we take representation and bias seriously. We will continue to do this for open ended prompts (images of a person walking a dog are universal!) Historical contexts have more nuance to them and we will further tune to accommodate that.”
He also responded to some critics directly by providing screenshots of his own interactions with Gemini which suggested the errors were not universal.
But the issues Gemini produced were quickly leveraged by anti-woke crusaders online, who claimed variously that Google was “racist” or “infected with the woke mind virus.”
Far-right internet troll Ian Miles Cheong blamed the entire situation on Krawczyk, whom he labeled a “woke, race-obsessed idiot” while referencing posts on X from years ago where Krawczyk acknowledged the existence of systemic racism and white privilege.
“We’ve now granted our demented lies superhuman intelligence,” Jordan Peterson wrote on his X account with a link to a story about the situation.
But the reality is that Gemini, or any similar generative AI system, does not possess “superhuman intelligence,” whatever that means. If anything, this situation demonstrates that the opposite is true.
As Marcus points out, Gemini could not differentiate between a historical request, such as asking to show the crew of Apollo 11, and a contemporary request, such as asking for images of current astronauts.
Historically, AI models including OpenAI’s Dall-E have been plagued with bias, showing non-white people when asked for images of prisoners, say, or exclusively white people when prompted to show CEOs. Gemini’s issues may not reflect model inflexibility, “but rather an overcompensation when it comes to the representation of diversity in Gemini,” says Sasha Luccioni, researcher at the AI startup Hugging Face. “Bias is really a spectrum, and it’s really hard to strike the right note while taking into account things like historical context.”
When combined with the limitations of AI models, that calibration can go especially awry. “Image generation models don’t actually have any notion of time,” says Luccioni, “so essentially any kind of diversification techniques that the creators of Gemini applied would be broadly applicable to any image generated by the model. I think that’s what we’re seeing here.”
As the nascent AI industry attempts to grapple with how to deal with bias, Luccioni says that finding the right balance in terms of representation and diversity will be difficult.
“I don’t think there’s a single right answer, and an ‘unbiased’ model doesn’t exist,” Luccioni said. “Different companies have taken different stances on this. It definitely looks funny, but it seems that Google has adopted a Bridgerton approach to image generation, and I think it’s kind of refreshing.”