A Reddit user posted two images made from the same prompt a day apart: “Amateur photograph of an elderly couple sat inside of a Yorkshire pub, amateur composition, candid.” The second looked less like AI image generation in the old sense and more like an actual bad pub photo someone might have pulled from Facebook in 2011. OpenAI’s verified products now make that shift harder to dismiss as a one-off: the newer ChatGPT image system and the API model gpt-image-1 are explicitly pitched around better prompt following, better text rendering, and easier everyday use. The important change is not that AI image generation became perfect. It is that plausible images are now cheap to generate and expensive to audit.
Commenters still found problems. The glasses were strange. The menu looked wrong. Chair geometry bent in impossible ways.
That used to mean the image failed.
Now it means the image is usable.
Why GPT Image 2 matters more than a prettier demo
“GPT Image 2” is an unverified Reddit nickname, not a confirmed OpenAI product name. But the reason that label spread is interesting: people were reacting to a capability jump in ordinary-looking photos, not just another polished demo.
The old way to judge image models was backwards. People asked whether the model could make something spectacular. A movie poster. A sci-fi city. A glossy ad.
The real test is whether it can make something boring.
A fake HOA complaint photo is more operationally useful than a gorgeous concept painting. A synthetic restaurant hygiene shot is more dangerous than a fantasy landscape. A made-up workplace incident image can do more reputational damage than a thousand beautifully rendered dragons.
Because those are the images that move through systems built for speed.
OpenAI’s official pages matter here not because they prove a Reddit nickname, but because they confirm the direction of travel. Better text rendering means signs and timestamps fail less often. Better prompt adherence means fewer retries. Easier image generation inside ChatGPT means the user no longer needs specialist tools or taste. A neighborhood agitator, a disgruntled customer, or a low-effort scammer can now produce plausible AI photographs with ordinary prompts and keep the best result.
That is the operational breakthrough.
Not “perfect fakes.” Fast enough fakes.
| Old giveaways | New giveaways | Why this survives feed-speed viewing |
|---|---|---|
| Mangled hands | Glasses hardware that doesn’t quite connect | Most people look at faces first, not hinge geometry |
| Melting faces | Reflections that don’t match the room | Reflections require zooming and cross-checking |
| Obvious gibberish text | Menus or signs that are almost readable but structurally wrong | “Almost right” text passes in a scroll |
| Hyper-stylized lighting | Chair legs, frames, and background geometry that bend subtly | Corners get less attention than subjects |
| Plastic stock-photo look | Repeated textures, warped edges, collapsed background details | Feed-speed viewing rewards gist, not inspection |
The old question was “does this look fake?”
The new one is “who has time to audit the corners?”
Realistic AI photos change the economics of trust
The big change in realistic AI photos is not visual. It is economic.
Generating candidates is cheap. Verifying them is slow.
That asymmetry is enough.
OpenAI’s API pricing and product packaging already make repeated trial generation normal. The exact per-image cost matters less than the behavior it enables: generate ten, keep one, crop the weak spot, repost, screenshot, circulate. The attacker gets retries. The verifier gets homework.
That changes what counts as “good enough.” A fake image no longer has to survive lab-grade analysis. It only has to survive one board member, one local reporter, one moderator, or one exhausted group chat.
Take the HOA example. The who, what, when, and why are simple: a neighbor wants to pressure a homeowners association over a dispute, so they post a “photo from this morning” showing overflowing trash, a broken fence, and junk piled in a yard. The first cues to inspect are not the central mess. Check the fence lines in the background, any mailbox numbers, shadows relative to the claimed time, and repeated debris textures near the edges. If those drift, the image probably drifted too.
Or the restaurant case. A customer claims a local restaurant mishandled food and posts a photo of raw chicken near produce on a stained prep surface. The fastest checks are timestamp typography, reflections in steel surfaces, glove edges, label text, and whether the cutting board wear pattern makes physical sense. A feed will read “gross kitchen.” An audit sees whether the room is coherent.
Or the workplace incident. An employee or troll posts an image of a factory spill, a missing safety guard, and an injured-looking hand near frame. Start with the guard’s attachment points, warning labels, floor reflections, hand posture, and any visible machine branding. Synthetic scenes often fail where equipment meets regulation. They know what an injury looks like. They are worse at what a compliant machine looks like.
None of these images need to be perfect.
They need to be believed for six hours.
That is enough to trigger a meeting, a call from a reporter, a reputational hit, or a moderation action. The effect comes first. The audit comes later.
This also strengthens the liar’s dividend. Once plausible fakes are common, real photos get easier to deny. “That was AI” becomes a usable defense even when it is false, because proving authenticity is work and doubt is cheap.
Cheap synthesis attacks both sides of trust: false accusations become easier to make, and genuine evidence becomes easier to wave away.
Here is the prediction. By Q1 2027, Meta will surface provenance status for newly uploaded images in at least one high-risk surface, most likely Facebook Groups, and will show a visible label or source-details panel on at least some uploads flagged as AI-generated or provenance-missing. Not because Meta suddenly loves epistemology. Because Facebook Groups are exactly where plausible local-evidence disputes become a moderation tax.
The real breakthrough is consistency, not just realism
People talk about photorealistic AI images as if the key question were whether they can fool an expert forever.
That is the wrong threshold.
The useful threshold is reliability under ordinary prompts.
If a normal user can type “photo of a pub couple,” “restaurant kitchen hygiene violation,” or “yard full of trash after storm” and get multiple plausible outputs without fighting the model, the system is already operationally useful. It does not need to be indistinguishable. It needs to be dependable enough.
That is why consistency matters more than peak quality. One astonishing image is a demo. A high hit rate is a workflow.
And workflows are what change behavior.
They change what people post. They change what scams look like. They change what moderators have to review. They change what a rumor can attach itself to.
The inspection pattern changes too. With older image systems, the main subject usually gave the game away. With newer ones, the image often passes the first glance and fails only under accounting.
The scene has bookkeeping errors:
– nose pads on glasses
– reflections in windows or frames
– menu layout and text hierarchy
– chair legs and table joins
– repeated textures in clutter
– shadows that disagree by a little
– labels and signage that are nearly right
Those are the new hands.
And they matter because they are slow. A seven-fingered hand shouts. A warped menu whispers. Most people never hear whispers in a feed.
This is also why the old question of whether people can spot AI-generated faces is now too narrow. Faces are only one part of the image. The background is where these systems increasingly confess.
There is a useful analogy to prompt injection: the danger is not theatrical failure but systems that are easy to manipulate under normal use. Image models are heading toward the same place. Not magical deception. Routine deception.
What readers can steal from the prompt-and-check workflow
The strongest practical response is to copy the workflow good image users already have.
Not “look harder.”
Look in the right order.
Verification box: the 60-second red-team check
1. Separate plausibility from evidence
A believable image is not proof anymore. Start there.
2. Check provenance before pixels
Who posted it first? Is there an original upload? Is there preserved metadata or C2PA information? Has it already been screenshotted through three platforms? Source chain beats vibes.
3. Audit the corners, not the subject
Inspect:
– glasses hardware
– reflections
– background text
– furniture geometry
– repeated textures
– edge collapse
– shadows that disagree slightly
4. Match the claim to the scene
Does the claimed time, place, and event fit the visible details? Weather, signage, uniforms, lighting, architecture, and equipment often betray synthetic scenes faster than faces do.
5. Look for sibling evidence
A real event usually leaves traces: another angle, another witness, another upload, a location match, a timestamp trail. A generated image often has only copies of itself.
Three red-team examples
Fake HOA complaint image
Claim: “Taken this morning. Unit 14 has trash everywhere again.”
Inspect first:
– house numbers and mailbox text
– fence alignment and property lines
– trash bag repetition
– shadow direction vs. claimed morning time
– parked cars or street signs that should identify the location
Fake restaurant hygiene photo
Claim: “Do not eat here. This was in the kitchen today.”
Inspect first:
– stainless-steel reflections
– food label text and date stamps
– glove seams and utensil edges
– cutting board wear pattern
– health-code signage or prep labels that should be standard in that locale
Fake workplace incident image
Claim: “Management ignored safety and someone got hurt.”
Inspect first:
– warning labels and machine branding
– bolt points where guards attach
– PPE consistency
– floor reflections and spill boundaries
– whether the injury, machine state, and scene layout physically match one another
A compact checklist helps:
| Step | What to do | What you’re looking for |
|---|---|---|
| Separate claim from proof | Treat the image as a claim attachment | Believability is not evidence |
| Provenance check | Find original upload and metadata | Source chain, C2PA, repost trail |
| Corner audit | Ignore the subject at first | Glasses, menus, reflections, geometry |
| Scene match | Compare image details to the stated event | Time, place, weather, equipment, signage |
| Corroboration | Search for sibling evidence | Other angles, witnesses, event trace |
For builders using an image generation model for harmless work like mockups or storyboards, the same process runs in reverse. Generate candidates. Audit the corners. Check text. Ask what factual claim the image appears to make. Then decide whether publishing it invites a misunderstanding you did not mean to create.
The people who get the most value from these tools will not be the people with magical prompts.
They will be the ones who can spot drift before someone else mistakes it for evidence.
Key Takeaways
- AI image generation has crossed an important threshold: the main failures are shifting from obvious weirdness to subtle forensic drift.
- The real change is economic, not aesthetic. Plausible images are cheap to generate and expensive to audit.
- Realistic AI photos become dangerous before they become perfect, because they only need to survive feed-speed belief.
- The best cues are often in the boring parts: glasses, reflections, menus, chair geometry, labels, and shadows.
- By Q1 2027, Meta is likely to expose provenance status on at least one high-risk upload surface because moderation costs will force the issue.
Further Reading
- The new ChatGPT Images is here, OpenAI’s rollout page for its newer ChatGPT image experience and how the company frames usability and scale.
- Introducing our latest image generation model in the API, OpenAI’s API announcement for gpt-image-1, including implementation details and pricing.
- Introducing 4o Image Generation, OpenAI’s earlier announcement highlighting prompt adherence, text rendering, and C2PA metadata.
