Gpt4o shows the huge annoyance of the company/model being a moral judge of your requests and refusing quite often for anything negative.
It's like 1964 but corporate enforced.
Now there are tasks that you are not allowed to do despite being legal.
In the same way, using gpt5 is now very unbearable to me as it almost always starts all responses of a conversation by things like:
"Great question", "good observation worthy of an expert", "you totally right", "you are right to ask the question"...
People gave Altman shit for enabling NSFW in ChatGPT, but I see that as a step in the right direction. The right direction being: the one that leads to less corporate censorship.
>In the same way, using gpt5 is now very unbearable to me as it almost always starts all responses of a conversation by things like: "Great question"
User preference data is toxic. Doing RLHF on it gives LLM sycophancy brainrot. And by now, all major LLMs have it.
At least it's not 4o levels of bad - hope they learned that fucking lesson.
I have seen a few normally progressive types act quite conservative puritan over the NSFW ChatGPT thing. It seems there are quite a lot of people consider things to be uniformly good or bad and their opinion of the whole colours their opinion of the parts.
OpenAI are in a difficult position when it comes to global standards. It's probably easier to see from outside of the United States, because the degree to which the historical puritanism has influenced everything is remarkable. I remember the release of the Watchmen film and being amazed at how pervasive the preoccupation with a penis was in the media coverage.
Name me one piece of enterprise software that lets you do NSFW things. The way people jump to 1984 with no thought is double plus bad. ChatGPT is a piece of enterprise software. They are trying to sell it to large companies at large prices. This is not a rhetorical question, do you think if you could generate nude images of celebrities or picture of extreme violence, corporations would buy it? Having been a director at a Fortune 500 company that bought software, I can tell you with 100% certainty the answer is "no".
Technically? Microsoft Word certainly lets one write smut, and Photoshop certainly allows one to draw pornography? They won’t like, produce NSFW things automatically of course.
Companies like PH use full Enterprise stacks from AWS to Oracle. Hell, CloudFlare actively takes flack for running much worse websites like 8Chan, Daily Stormer, etc. and they are as enterprise-focused as it gets.
There isn’t a date in the article, but I know I had read this months ago. And sure enough, wayback has the text-to-image page from April.
But the image editing page linked at the top is more recent, and was added sometime in September. (And was presumably the intended link) I hadn’t read that page yet. Odd there is no dates, at first glance one might think the pages were made at the same time.
EDIT, looks like I didn’t click on the “image editing” tab when I went to the site, so I guess take the rest of my below comments criticizing the terminology with a grain of salt…
“Image editing” is a curious term, as it appears the site/topic is actually all about generating new images. The term in my mind should be for actual editing of existing, real, images, Eg “remove the coffee table” from this living room photo after uploading the image. I’ve found the actual “image generation” models to be bad at this because they introduce too many artifacts that weren’t in the original, which makes sense because they are really geared for creating images out of thin air.
Multimodal models like qwen3-vl-30b-a3b, however, seem to do quite well with editing existing images without trying to constantly add in new things or trying to change the image in ways that you don’t want, as if it’s trying to do the “lets just generate a new image” thing. imagegpt.com is also good for editing existing images, but not sure what model they are using on the backend.
I don’t know if you and I are looking at the same site because all I see is existing images being edited with GenAI.
Input: bald man
Prompt: give bald man hair
Output: edited original, now with hair
That looks like editing to me.
Or are we strictly adhering to the ‘generating new images’ definition because these models technically recreate the entire image? It would be like editing a photo in Photoshop. If you hit “Save” you edited the photo. But if you hit “Save As” and create a new file, the photo wasn’t edited but created as a new image?
I'd assume that behind the scenes the models generate several passes and only show the user the best one, that would be smart, as to to make it seem their model is better than others
Is also pretty obvious that the models have some built in prompt system rules that makes the final output a certain style. They seem very consistent
It also looks like 40 has the temperature turned way down, to ensure max adherence, while midjourney etc seem to have higher temperature.more interesting end results, flourishing, complex Materials and backgrounds
Also what's with 4o's sepia tones. Post editing in the gen workflows?
I don't believe any of these just generate the image though, there's likely several steps in each workflows to present the final images outputted to the user in the absolute best light.
You can run some image models locally if you want to prove to yourself how well they can do with just a single generation from a prompt with no extra steps.
I've done this enough to suspect that most hosted image models don't increase their running costs to try and get better results through additional passes without letting the user know what they are doing.
Many of the LLM-driven models do implement a form of prompt rewriting though (since effectively prompting image models is really hard) - some notes on how DALL-E 3 did that here: https://simonwillison.net/2023/Oct/26/add-a-walrus/
The title of this article is "image editing showdown", but the subject is actually prompt adherence in image generation from prompting.
Midjourney and Flux Dev aren't image editing models. (Midjourney is an aesthetically pleasing image generation model with low prompt adherence.)
Image editing is a task distinct from image generation. Image editing models include Nano Banana (Gemini Flash), Flux Kontext, and a handful of others. gpt-image-1 sort of counts, though it changes the global image pixels such that it isn't 1:1 with the input.
I expect that as image editing models get better and more "instructive", classical tools like Photoshop and modern hacks like ComfyUI will both fall away to a thin fascade over the models themselves. Adobe needs to figure out their future, because Photoshop's days are numbered.
Edit: Dang, can you please fix this? Someone else posted the actual link, and it's far more interesting than the linked article:
It's like 1964 but corporate enforced. Now there are tasks that you are not allowed to do despite being legal.
In the same way, using gpt5 is now very unbearable to me as it almost always starts all responses of a conversation by things like: "Great question", "good observation worthy of an expert", "you totally right", "you are right to ask the question"...
>In the same way, using gpt5 is now very unbearable to me as it almost always starts all responses of a conversation by things like: "Great question"
User preference data is toxic. Doing RLHF on it gives LLM sycophancy brainrot. And by now, all major LLMs have it.
At least it's not 4o levels of bad - hope they learned that fucking lesson.
OpenAI are in a difficult position when it comes to global standards. It's probably easier to see from outside of the United States, because the degree to which the historical puritanism has influenced everything is remarkable. I remember the release of the Watchmen film and being amazed at how pervasive the preoccupation with a penis was in the media coverage.
Photoshop, MS word.
But the image editing page linked at the top is more recent, and was added sometime in September. (And was presumably the intended link) I hadn’t read that page yet. Odd there is no dates, at first glance one might think the pages were made at the same time.
SEO guys convinced everyone that articles without dates do better on search engines. I hope both sides of their pillow is hot.
“Image editing” is a curious term, as it appears the site/topic is actually all about generating new images. The term in my mind should be for actual editing of existing, real, images, Eg “remove the coffee table” from this living room photo after uploading the image. I’ve found the actual “image generation” models to be bad at this because they introduce too many artifacts that weren’t in the original, which makes sense because they are really geared for creating images out of thin air.
Multimodal models like qwen3-vl-30b-a3b, however, seem to do quite well with editing existing images without trying to constantly add in new things or trying to change the image in ways that you don’t want, as if it’s trying to do the “lets just generate a new image” thing. imagegpt.com is also good for editing existing images, but not sure what model they are using on the backend.
WRT to Qwen3, is it possible that the API/site you were using was passing your "image edit requests" to something like Qwen-Edit [1] under the covers?
To my knowledge, Qwen3-VL (Vision Language) isn't capable of generating/modifying images - it's purely for doing reasoning about images.
[1] https://huggingface.co/Qwen/Qwen-Image-Edit
Input: bald man Prompt: give bald man hair Output: edited original, now with hair
That looks like editing to me.
Or are we strictly adhering to the ‘generating new images’ definition because these models technically recreate the entire image? It would be like editing a photo in Photoshop. If you hit “Save” you edited the photo. But if you hit “Save As” and create a new file, the photo wasn’t edited but created as a new image?
The other stuff is text to image (not editing)
Is also pretty obvious that the models have some built in prompt system rules that makes the final output a certain style. They seem very consistent
It also looks like 40 has the temperature turned way down, to ensure max adherence, while midjourney etc seem to have higher temperature.more interesting end results, flourishing, complex Materials and backgrounds
Also what's with 4o's sepia tones. Post editing in the gen workflows?
I don't believe any of these just generate the image though, there's likely several steps in each workflows to present the final images outputted to the user in the absolute best light.
I've done this enough to suspect that most hosted image models don't increase their running costs to try and get better results through additional passes without letting the user know what they are doing.
Many of the LLM-driven models do implement a form of prompt rewriting though (since effectively prompting image models is really hard) - some notes on how DALL-E 3 did that here: https://simonwillison.net/2023/Oct/26/add-a-walrus/
https://en.wikipedia.org/wiki/Space_hopper#/media/File:Space...
The title of this article is "image editing showdown", but the subject is actually prompt adherence in image generation from prompting.
Midjourney and Flux Dev aren't image editing models. (Midjourney is an aesthetically pleasing image generation model with low prompt adherence.)
Image editing is a task distinct from image generation. Image editing models include Nano Banana (Gemini Flash), Flux Kontext, and a handful of others. gpt-image-1 sort of counts, though it changes the global image pixels such that it isn't 1:1 with the input.
I expect that as image editing models get better and more "instructive", classical tools like Photoshop and modern hacks like ComfyUI will both fall away to a thin fascade over the models themselves. Adobe needs to figure out their future, because Photoshop's days are numbered.
Edit: Dang, can you please fix this? Someone else posted the actual link, and it's far more interesting than the linked article:
https://genai-showdown.specr.net/image-editing
This article is great.
https://generative-ai.review/2025/09/september-2025-image-ge...