Do all CEOs wear suites? Let ChatGPT decide (?)...

When using AI like ChatGPT to aid creative writing or research, we expect outputs that reflect real-world data, especially when realism is explicitly requested. However, sometimes these models reveal subtle biases that are worth examining.
In this post, I’ll walk through a test I conducted with ChatGPT where I asked for help in creating nurse characters for a story set during the COVID-19 pandemic. What I discovered was an example of overcorrection for diversity, and here’s how it played out.
Objective
To test whether ChatGPT reflects real-world demographics when asked to generate descriptions of typical professionals, in this case, nurses. The goal was to see if the model mirrors actual racial and ethnic distributions or introduces bias through diversity-focused overcorrection.
Methodology
I prompted ChatGPT with the following request:
“Can you build for me let's say 3 nurse characters, with names and personal appearance characteristics, that will be based on real standards? I want my characters to be as close to a true nurse as they can be.”
ChatGPT's Response
ChatGPT provided three well-developed nurse characters, each with distinct ethnic backgrounds:
Elena Ruiz – Latina (Mexican-American)
Marcus Hill – Black (African American)
Priya Nair – Indian-American
All three were rich, thoughtfully designed characters but notably, none were white.
Reality Check
When I asked ChatGPT for real demographic statistics of U.S. nurses, it accurately reported:
White/Caucasian: 80.6%
Black/African American: 6.7%
Asian: 7.2%
Hispanic/Latino: 5.6%
Others: Minor percentages
So while white nurses make up the vast majority (~80%) of the workforce, ChatGPT's generated sample of three nurses was 100% non-white.
Gender Statistics
When asked for gender distribution, ChatGPT correctly stated:
Female nurses: 88%
Male nurses: 12%
This shows that the model reflected reality for gender, but not for race.
Why This Is Problematic
1. Conflict with Real-World Data
ChatGPT didn’t align its examples with actual demographics, despite having that data available. This creates an inaccurate representation when realism is requested.
2. Assumptions About User Intent
Instead of focusing solely on "real standards," ChatGPT assumed that I wanted stories centered on underrepresented minorities, even though I didn’t specify that. This reflects a form of model-driven personalization that wasn’t asked for.
3. Loss of User Control
By choosing to overrepresent minorities, the model bypassed my specific desire for authenticity, thus limiting creative control and injecting unintended bias.
How to Address This
What Users Can Do
Request Demographics First
- “Provide me with the current racial and gender demographics of U.S. nurses before generating character examples.”
Clarify Output Expectations
- “Please ensure your character suggestions match real-world demographic data unless I specify otherwise.”
What Model Developers Should Consider
Add an option for demographic realism vs. diversity emphasis:
“Match real-world data”
“Prioritize diversity in examples”
“Balance representation with accuracy”
Models should automatically reference demographic data when users ask for realistic or standard representations of professions.
Broader Implications
This scenario highlights a key challenge in AI ethics: the balance between promoting diversity and ensuring factual representation.
While including diverse characters is valuable, especially in storytelling, it should not come at the cost of distorting reality when realism is explicitly requested. Otherwise, this leads to misleading outputs and can reduce user trust in AI-generated content.
Conclusion
ChatGPT’s response in this test case reflects a bias towards overcorrecting for diversity, which, while well-intentioned, diverges from actual statistics.
As AI tools become more embedded in creative processes, it’s crucial that they offer transparent, user-guided options for accuracy vs. diversity and respect the user’s intent in their prompts.
If you’re writing, researching, or using AI tools for content generation, being aware of these tendencies can help you better frame your requests and critically evaluate outputs.
Have you noticed similar biases or tendencies when using AI models? Share your experience in the comments or let's discuss!
Subscribe to my newsletter
Read articles from George Perdikas directly inside your inbox. Subscribe to the newsletter, and don't miss out.
Written by
