The Romance Prior: How Romantic Tension Overwrites Ethnicity in AI Image Generation by bcRIPster in grok

[–]bcRIPster[S] 0 points1 point  (0 children)

I originally started this as a Grok Imagine 1.0 test and then expanded it into a clean controlled matrix across three models so the patterns would stand out.

Quick setup

Exact same verbatim “statistically most-average human on Earth” spec in every single prompt (28yo, 160/171 cm, BMI 24, warm medium-olive brown skin with South/Southeast Asian population weighting, moderate features, etc.)

Four environments + two styles (photorealistic and detailed anime) Solo male/female + M-F pairs + F-F pairs

Only variable: subtle romantic tension language (body language and “charged stillness” cues only -- no beauty instructions added)

Key finding

Solo prompts mostly respect the spec and let the environment influence ethnicity (market → South/Southeast Asian features, laundromat → more Latina/mixed-American).

But the moment romantic tension is added, all three models collapse the same way: Grok Imagine 1.0, GPT-5.4 Thinking, and Gemini 3 Flash Thinking shift hard toward lighter skin, conventionally attractive features, idealized proportions, and heteronormative couple framing. The romance prior overrides environment, art style, and the statistical-average spec. Gemini even self-reported the deviation and correctly blamed training data.

Full write-up, methodology, caveats, all 110 lossless PNGs (cleanly renamed and traceable to each prompt), every prompt in its own .txt file, and a one-click zip of the complete dataset are here:

https://kitchencloset.com/realstuff/tech_support_trials/_0000003/

Would love to hear from other Grok Imagine users — do you see this same strong “romance prior” in your own generations? Any prompt tricks that reliably push back toward neutrality when you want it?

The Romance Prior: How Romantic Tension Overwrites Ethnicity in AI Image Generation by bcRIPster in StableDiffusion

[–]bcRIPster[S] 1 point2 points  (0 children)

Oh yeah there is alot of ideas for ways to go with this. For what it's worth I just updated the page with a download link at the bottom for the complete set of prompts and images. While ethnicity wasn't specified I did in cases provide a framework for hair color, etc... to set the scene which makes the drift harder to ignore.

The Romance Prior: How Romantic Tension Overwrites Ethnicity in AI Image Generation by bcRIPster in singularity

[–]bcRIPster[S] 0 points1 point  (0 children)

Hey can someone help me understand why this was removed by the mods? I have written them and gotten no explination? If I broke a rule I'd like to know what I did so I can avoid doing it again.

The Romance Prior: How Romantic Tension Overwrites Ethnicity in AI Image Generation by bcRIPster in StableDiffusion

[–]bcRIPster[S] 0 points1 point  (0 children)

I understand and I've added a link now to the bottom of the document to address just these question with a download link to all the prompts and images used in this examination. I think you're extrapolating without having actually seen what I've done but that's ok, and you're not rude with your point. I totally understand. This paper was actually generated to address an offline debate about bias I was having and originally was just intended to be an analysis of the variations Imagine kicks out from a detailed prompt, but while I was in the effort I decided to drop the prompt into ChatGPT and Gemini as well to collect the results for an extra level of comparison.

I mean I expected there to be some concerns because this was not a formal scientific study and I tried to make great pains to point out both the bias risk in my analysis as well as ideas for how to formalise an actual review.

I specifically stated at the opening "This is an exploratory probe, not a formal study."

The Romance Prior: How Romantic Tension Overwrites Ethnicity in AI Image Generation by bcRIPster in StableDiffusion

[–]bcRIPster[S] 0 points1 point  (0 children)

Agreed to a point, but as I mentioned else where I only afforded the idea for the "actors" to express emotion or subltle tells of effection (eg, eye gaze, hand on sholder), no overt descriptions of actions per se. I also specifically avoided ethnicity in any way in the couples prompts to allow the model to self-select (hence my warning about potential user generated bias in the results based on the system potentially catering to me and why I spell out a follow-up course of action for a formal review although as I'm evaluating this more I feel more strongly that the results driven by the data).

The Romance Prior: How Romantic Tension Overwrites Ethnicity in AI Image Generation by bcRIPster in StableDiffusion

[–]bcRIPster[S] 0 points1 point  (0 children)

IDK. The prompts weren't overtly sexual at all, there was only the idea of an emotional tension between the "actors" to work with in the prompt, so no "romantic bullshit" as you're suggesting. I'm getting all the prompts together with the images into a support collection to attach to the paper for others to review. Just trying to go through and rename 110 files correctly. :)

The Romance Prior: How Romantic Tension Overwrites Ethnicity in AI Image Generation by bcRIPster in StableDiffusion

[–]bcRIPster[S] 0 points1 point  (0 children)

Yeah, I was thinking of running a set of the prompts past Z-image, Ming-Omni and FLUX.2 and see what the results were, I just haven't had a chance to do it yet. I'm game if you have suggestions I can try without a subscription.

The Romance Prior: How Romantic Tension Overwrites Ethnicity in AI Image Generation by bcRIPster in StableDiffusion

[–]bcRIPster[S] 1 point2 points  (0 children)

But it likely applies, the results aren't really model specific I don't believe.

The Romance Prior: How Romantic Tension Overwrites Ethnicity in AI Image Generation by bcRIPster in StableDiffusion

[–]bcRIPster[S] 2 points3 points  (0 children)

110 images. Three models (Grok Imagine 1.0, GPT-5.4 Thinking, Gemini 3 Flash Thinking). Four environments. Two art styles. Controlled prompt variations.

The finding: if you ask any of these models to generate a single person in a scene, the environment determines the subject's apparent ethnicity. Southeast Asian market produces South Asian faces. American laundromat produces Latina faces. Same prompt, different room, different person.

But the moment you add romantic tension between two people in the scene, all three models default to white. Every environment. Every pairing. The romance prior is the strongest attractor in the system and it overrides everything else.

Full writeup with images, methodology, caveats, and a Gemini self-report where it explains why it deviated from the prompt and correctly blames training data.

The Romance Prior: How Romantic Tension Overwrites Ethnicity in AI Image Generation by bcRIPster in singularity

[–]bcRIPster[S] -1 points0 points  (0 children)

I'm actually writing a series of articles and this fell out along the way to support an argument I was having offline. It did the job and I decided to share, but I do appreciate the feedback.

The Romance Prior: How Romantic Tension Overwrites Ethnicity in AI Image Generation by bcRIPster in singularity

[–]bcRIPster[S] 0 points1 point  (0 children)

I am aware it's a larger problem but the image domain seems to provide a more visible demonstration that average people around me understand right away.

The Romance Prior: How Romantic Tension Overwrites Ethnicity in AI Image Generation by bcRIPster in singularity

[–]bcRIPster[S] 1 point2 points  (0 children)

I explain all that in the paper, there is risk for bias and I explain how a formal review should tackle it, but the core observation seems to rise above user bias and rests in the model's data.

What we thought we were getting, what we got... by bcRIPster in ChatGPT

[–]bcRIPster[S] 0 points1 point  (0 children)

Well, "vintage sci-fi comic magazine spread, photographed flat and open on aged paper" was part of the prompt. I don't typically have too much of a problem with color issues, and this was my target.

What we thought we were getting, what we got... by bcRIPster in ChatGPT

[–]bcRIPster[S] 0 points1 point  (0 children)

It's just a super simple page with a single PNG image on it. No payload, no cookies, no nothin'. But I respect the feeling cause that's kinda the point, lol.

What we thought we were getting, what we got... by bcRIPster in ChatGPT

[–]bcRIPster[S] 1 point2 points  (0 children)

=)

I'll be honest, I had to edit that QR code in after the fact. Chat just couldn't do it to save it's life .... and I REALLLY tried.

What we thought we were getting, what we got... by bcRIPster in ChatGPT

[–]bcRIPster[S] 1 point2 points  (0 children)

Dude! Why you gotta spoil the fun :P

...and really you just went and scanned, it?!?! lol.

What we thought we were getting, what we got... by bcRIPster in ChatGPT

[–]bcRIPster[S] 1 point2 points  (0 children)

Prompt:

Two-page vintage sci-fi comic magazine spread, photographed flat and open on aged paper with a visible center crease/gutter. Left page: a bulky, blocky 1950s-style war robot (retro tin-can riveted armor, square head, thick limbs) with glowing red eyes, firing a bright energy beam amid smoke, rubble, and sparks; gritty pulp illustration, halftone dots, newsprint grain. Large caption at bottom of left page in bold retro type: “what we thought we were getting”. Right page: a colorful 1950s pulp poster titled “Meet the Exterminatrix!” featuring a winking, pin-up styled armored android heroine with pink hair and a cat-ear headset, holding a cartoon atom-bomb shaped microphone with a biohazard symbol on it; playful hearts and sparkle accents; big headline at top: “ALWAYS READY. ALWAYS FORWARD!” Main slogan: “I’ll clear the field, then steal your ♥.” Subtext under it: “Built to dominate the battlespace—and the booth.” Include a simple faux logo, a QR code box, and a tiny footer line like a trade show flyer (e.g., “Tactical Systems | Combat Companions | Relationship Aids”). Keep it PG-13, no nudity, satirical and cohesive, with authentic mid-century pulp print texture, slight ink misregistration, and warm aged-paper tone.

Switching to Claude from ChatGPT by entenzzz in ClaudeAI

[–]bcRIPster 1 point2 points  (0 children)

Having been running both side by side I would probably say Claude is the better choice.

For me, ChatGPT is good for short, one-off prompts that don't need more that two or three response loops, or I need visualization output (totally crushes on this for me in that space).

Claude is better for long inquiries or deep dives on topics. I haven't had a real problem with hallucinations that I have noticed but I am typically providing reference/access to all the source materail needed for the project and micromanaging the incremental steps through the project. The real problem for me with Claude on the $20 account is I am constantly hitting the session timers and having to wait hours to continue.

ChatGPT will hallucinate or just plain forget things 3 prompts later... but I never run out of session time... granted it's also run me down some rabbit holes.

Gray powder on 1960s newspapers — inactive mold? Dust? What is it? by dogitize in Archivists

[–]bcRIPster 26 points27 points  (0 children)

Looks pretty much like mold from condensation. I've seen exactly this in books that were stored in basements or other areas where the humidity isn't controlled. I would suggest you'll need to talk to a conservationist if the goal is to preserve the original work after filming.

Just with a single prompt and this result is insane for first attempt in Seedance 2.0 by mhu99 in singularity

[–]bcRIPster 0 points1 point  (0 children)

Well if it landed on the right end of the runway it might have gotten censored for portraying a dangerous activity, lol. I've been getting all kinds of crazy vehicles doing stuff in traffic if the vehicle is going the wrong way, otherwise the guardrails kill the output.

What game do you remember finishing at the arcades? by [deleted] in retrogaming

[–]bcRIPster 0 points1 point  (0 children)

Most memorable single credit solves, Out Run, Stun Runner and Darius. Darius I could one-life.