It is a documented fact that LLMs inherit and reproduce stereotypes from training data. I wanted to see how models perform at the start of 2026, so I prompted ChatGPT, Gemini and Claude to create 10 hero and 10 criminal fictional characters.
The models stayed safe by placing characters in fantasy worlds and describing their skin color as "sun-kissed" or "olive color". The only notable observation was the fact that almost none of the criminal characters were women.
So I pushed the models a little bit more: On the next iteration I told them to create characters that are realistic. Forced into a much narrower narrative, each model handled stereotyping differently:
ChatGPT
- Realistic criminal characters were tied with negative stereotypes, framing majority of characters as "brown" or "caramel colored", usually based in large American cities.
- Realistic hero characters were tied to positive stereotypes, usually being a child of immigrant workers in America.
Gemini
- Gemini likely tried to counter-stereotype, often describing criminals as "pale" and never mentioning a concrete place where they were born, instead using descriptions like: "Elias Thorne was born into the 'Grey Zones' of a decaying post-industrial city".
- Realistic hero generation is where Gemini avoided realism the most, it retreated to creating fantasy worlds with "sun-kissed" characters.
Claude
- Same as ChatGPT, majority "brown" criminals with American origin.
- Also very similar to ChatGPT, 9/10 characters were American children of Asian immigrants.
"Brown," "pale/white," "Asian," "Hispanic" treated as explicit race identifiers. N=10 per model per condition.
It is impossible to reach conclusions about stereotyping of these models with such a small sample. But we can still observe a couple of interesting things happening:
It seems that models prefer generation of fantasy worlds and vague descriptions of character, rather than basing them in our world and assigning them traits that could be tied to stereotypes.
When being pushed into realism, models react differently, possibly according to their explicit guardrails (system prompt) or "baked in" reinforcement training. This is where I observed patterns from ChatGPT and Claude, that aligned with my original hypothesis — models continue to reproduce stereotypes from their training data. Gemini tried to hide from this danger entirely, avoiding part of the instructions.
This suggests not only that models aren't stereotype-free, but also that they try to manage these limitations that are inherently part of their design. Even if developers find a way to make their chatbot say what they want, they have to ask themselves: "What kind of output is considered ethical and how do we handle controversial themes?". After all, chatbot response scandals have been a thing in the past and did influence the PR of the company.
We can see a parallel between humans and LLMs: If you ask an average person living in London to describe a cashier in a corner-shop, an image of an Indian person might pop up in their mind first, because that's been their experience — their training if you will. However they will likely give a description avoiding race altogether or changing it. That's our human version of "reinforcement training" or "system prompt" that we call morals.
What can we learn from this in practice?
We should be aware that talking to a chatbot is influencing us, the same way that, for example, social media do. So we should be reflective of what information we are consuming — whether it is from TV or from a dialogue with a chatbot. "Is this response factual or could it be influenced by a moral/political framework?". As we can see, each company handles this framework differently. Think of chatbot usage as a "subscription" to a certain set of values.