The user suggests that instead of forcing a switch to a new model like GPT Image 1.5, platforms should allow it as an optional or fallback model. This enables users to test where it excels (e.g., iteration, edits, edge cases) and promotes a 'model-agnostic' approach, ensuring users can leverage the best model for their specific core use cases.
BREAKING NEWS: OpenAI launched their updated image generation model called GPT Image 1.5. It's a competitor to the much lauded Nano Banana Pro model by Google. I'll be honest. Nano Banana Pro is so good for the kinds of things I use image generation for that I didn't think GPT Image 1.5 was going to be able to match it. Specifically, Nano Banana can reason over a detailed prompt and create very detailed images (like infographics) that include lots of text. I figured the new OpenAI model would be great at photo realistic images and maybe some text (like for ads and such) and *maybe* decent at editing/iterating on images, but I didn't think it would be able to create business graphics in the way Nano Banana does. Which is why I use it in my image generation agent (ImageGen .ai). But, I was wrong. I generated an image with this simple prompt: "infographic of major milestones for HubSpot". Below is the image generated in one shot (no iteration). It is exceptional and right on par with what Nano Banana Pro can do. * It actually used the HubSpot logo in the image (with no reference image provided) * It included cartoon images of me and Brian Halligan. * It got all the text right. And there's a lot of it. * It used a color scheme that is on brand (without specific prompting). * I'm no designer, but it looks pretty good. A couple of nits with it (like the way the timeline flows), but overall, I think it did a stellar job. Now the question I'm pondering: Should I be using this model in my ImageGen .ai agent (instead of Nano Banana Pro)? I'm not sure. Right now, I'm leaning against, as I don't think it's dramatically better -- just equally good. What do you think?