🚀 OpenAI's Secret "Image V2" Just Leaked on LM Arena: The End of Mangled AI Text?

Dev.to / 4/8/2026

💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • A report from TestingCatalog claims OpenAI is running a stealth A/B test for a next-generation image model internally dubbed “Image V2,” with variants allegedly appearing on LM Arena before being pulled.
  • Early user reports suggest Image V2 markedly improves realistic UI rendering and produces correctly spelled text, addressing a long-standing failure mode in AI-generated marketing and interface assets.
  • The article links the initiative to “code red” pressure within OpenAI since late 2025, implying urgency driven by competitive dynamics with Google.
  • If the leak reflects a real update, designers and frontend developers could generate more reliable text-accurate landing page mockups and reduce manual correction work before implementing code.
  • The testing reportedly mirrors a prior OpenAI blind-testing approach used in December 2025 with “Chestnut” and “Hazelnut,” which later shipped as GPT Image 1.5.

If you've been using ChatGPT over the weekend and suddenly found yourself being asked to choose between two surprisingly high-quality image generations, congratulations—you might be an unwitting beta tester for OpenAI’s next major release.

According to a new report from TestingCatalog, OpenAI is quietly running a massive stealth test for its next-generation image generation model, internally dubbed "Image V2." If you build apps, design UIs, or generate commercial assets, this is a massive deal. Here is everything we know about the leaked model, the "code red" pressure from Google, and why this update might finally fix AI's biggest, most annoying flaw. 👇

🕵️‍♂️ The Arena Leak: What is "Image V2"?

Over the past few days, eagle-eyed users on the LM Arena (the premier blind-testing leaderboard for AI models) noticed three mysterious new image generation variants pop up:

  • packingtape-alpha
  • maskingtape-alpha
  • gaffertape-alpha

By the end of the weekend, the models were pulled from the Arena, but they are still heavily circulating inside ChatGPT under a strict A/B testing framework.

This is classic OpenAI. They used this exact same blind-testing playbook back in December 2025 with the "Chestnut" and "Hazelnut" models, which ended up shipping just weeks later as GPT Image 1.5.

🤯 The Holy Grail: AI That Can Actually Spell

So, why should developers and designers care? Because early impressions indicate that Image V2 has finally conquered the final boss of AI image generation: Realistic UI rendering and correctly spelled text.

Historically, asking an AI to generate a UI mockup or a marketing banner resulted in beautiful designs covered in alien hieroglyphics. Image V2 is reportedly delivering pixel-perfect button text, accurate typography, and an incredibly strong compositional understanding.

If you are a frontend developer, this means you can soon prompt ChatGPT to generate a complete, text-accurate landing page mockup, slice it up, and start coding—without having to mentally translate mangled letters.

🚨 The "Code Red" Counter-Attack

It's no secret that OpenAI has been feeling the heat. According to the report, OpenAI has been operating under a CEO-mandated "code red" since late 2025.

Why? Because Google's Nano Banana Pro and Gemini 3 models have been absolutely eating their lunch, dominating the top spots on the LM Arena leaderboard for months. Image V2 is OpenAI’s direct, aggressive answer to Google's visual dominance.

💻 What the API Might Look Like

While pricing and official release dates are still unannounced, history shows OpenAI usually drops the new models into their existing SDK within weeks of these Arena tests. GPT Image 1.5 already slashed API costs by 20%, so we are hoping for competitive pricing here.

When it does drop, integrating the new model into your Node.js apps will likely be a seamless drop-in replacement. Here is how you'll probably trigger the new high-fidelity UI generations:

import OpenAI from "openai";

const openai = new OpenAI({ apiKey: process.env.OPENAI_API_KEY });

async function generateUIMockup(promptText) {
  console.log("🎨 Generating UI Mockup with Image V2...");

  const response = await openai.images.generate({
    model: "image-v2", // 👈 The anticipated new model name
    prompt: `A modern, clean SaaS dashboard UI. 
             Sidebar on the left with navigation. 
             Main content shows a revenue chart. 
             A bright blue button in the top right that explicitly says "Export Data". 
             High fidelity, web design, vector style.`,
    n: 1,
    size: "1024x1024",
    quality: "hd", // Requesting maximum text clarity
  });

  const imageUrl = response.data[0].url;
  console.log(`✅ Success! View your mockup here: ${imageUrl}`);
  return imageUrl;
}

generateUIMockup();

🔮 The Verdict

The biggest question now is whether OpenAI will maintain the incredible raw quality seen in the Arena, or if they will dial it back with heavy safety filters and cost-optimizations before the public API launch.

Either way, the era of AI failing to spell basic words on a button is coming to an end.

Have you encountered any of the "tape-alpha" models in your ChatGPT sessions this week? Did the text actually make sense? Let me know what you generated in the comments below! 👇

If you found this breakdown helpful, drop a ❤️ and bookmark this post so you're ready to update your API calls the minute the model officially drops!