ChatGPT Images 2.0 is here! Text generation accuracy is greatly improved, making it easy to create marketing posters

ChainNewsAbmedia

OpenAI officially released ChatGPT Images 2.0 on Tuesday, significantly improving the accuracy of text generation, as well as the design aesthetics for posters and portraits. The model also introduced “thinking mode” for the first time, enabling image generation to include web search and multi-image batch output capabilities, bringing it fully in line with commercial application scenarios.

(Canva announces deep integration with Claude, enabling the transformation of AI drafts into finished design deliverables)

From making things up to perfect menus: AI has finally learned to spell

Looking back two years ago, the weaknesses of AI image generation models in text generation were almost universally known. As long as the prompt included text requirements, the output would often be filled with absurd spelling errors or even hallucinations. This was even worse in non-English Chinese, Japanese, and Korean languages.

Official announcement Korean poster mockup

Now, ChatGPT Images 2.0 can generate a promotional poster that can be used directly by vendors, with clear and accurate text. In recent years, researchers have actively explored new architectures such as (Autoregressive Models) that can self-regress, and its operating logic, understanding of text, and generation and verification capabilities have improved significantly.

Thinking mode goes live: web search and composition consistency are all covered

The most core upgrade in ChatGPT Images 2.0 is “thinking mode (Thinking Capabilities).” It is currently available to paid users of ChatGPT Plus, Pro, the business version, and the enterprise version. Once enabled, the model can instantly perform web searches to assist image generation. It can also create corresponding visual explanation graphics based on files the user uploads, and conduct self-review and optimization of the image content before official output.

In batch generation, under thinking mode, a single prompt can output up to eight images at once, and the images can maintain consistent character appearances, object styles, and overall visual style. This makes it suitable for comic panels, social media series image-and-text posts, and even interior design space planning drawings for various rooms.

Official announcement comic panel mockup

In terms of resolution, the new model supports up to 2K output, and also adds multiple aspect ratio options from 3:1 to 1:3, further meeting a variety of business needs.

Asian languages are greatly optimized—Chinese, Japanese, and Korean users are in luck!

Besides English, OpenAI specifically noted major improvements to Images 2.0 for Asian text, including clear enhancements in Japanese, Korean, and Chinese.

Test articles that circulated in Chinese tech communities a few days ago also verified the news. Multiple Zhihu creators conducted hands-on testing comparisons between GPT-Image-2 and the competing Google Nano Banana Pro, covering a range of scenarios such as Chinese poster design, e-commerce cover images, social media interface layouts, and data visual charts.

Zhihu article tests GPT-Image 2.0

The test results show that GPT-Image-2 clearly outperforms in the aesthetics of Chinese typefaces, layout hierarchy, and overall design feel. The generated poster styles are closer to real commercial materials, rather than template-like outputs with an obvious “AI look.” The article also points out that GPT-Image-2 shows higher detail accuracy in replicating the interface (such as game screenshots or messaging app screenshots) and in recreating real portrait scenes.

ChatGPT Images 2.0 fully opens up, and the API also launches

Currently, ChatGPT Images 2.0 has been providing basic functionality to all ChatGPT and Codex users free of charge starting this Tuesday, while paid users can unlock more advanced output effects. At the same time, OpenAI is also opening the GPT-Image-2 API. Pricing is calculated based on output quality and resolution tiers, offering integration flexibility for enterprises and developers.

It’s worth noting that the new model’s knowledge cutoff date is December 2025. For image generation prompts involving the latest current events, accuracy may be subject to certain limitations. In addition, the generation speed for complex compositions can’t be as immediate as typical text Q&A, but it still only takes a few minutes.

This article: ChatGPT Images 2.0 makes its debut! Text generation accuracy greatly improves, making it easy to produce marketing posters First appeared on Chain News ABMedia.

Disclaimer: The information on this page may come from third parties and does not represent the views or opinions of Gate. The content displayed on this page is for reference only and does not constitute any financial, investment, or legal advice. Gate does not guarantee the accuracy or completeness of the information and shall not be liable for any losses arising from the use of this information. Virtual asset investments carry high risks and are subject to significant price volatility. You may lose all of your invested principal. Please fully understand the relevant risks and make prudent decisions based on your own financial situation and risk tolerance. For details, please refer to Disclaimer.

Related Articles

OpenClaw, Hermes, and SillyTavern Confirmed in GLM Coding Plan Support

Zhipu AI PM Li announces OpenClaw, Hermes, and SillyTavern as supported GLM Coding Plan projects; other tools will be evaluated case-by-case. Do not share credentials or use subscriptions as API access; contact support for error 1313. Zhipu AI product manager Li announced that OpenClaw, Hermes, and SillyTavern are officially supported under the GLM Coding Plan, with other tools evaluated case-by-case. The note cautions against sharing credentials or using subscriptions as API access and directs users with error 1313 to contact support.

GateNews4m ago

Google Cloud CEO: Gemini to Power Apple's Personalized Siri Rollout in 2026

Summary: Gemini will power a personalized Apple Siri in 2026, built on Apple's Foundation Models and Gemini collaboration; Apple tests a chat-like Siri in iOS 27/macOS 27, slated for WWDC 2026. Abstract: Google Cloud's Gemini is set to power a personalized Apple Siri by 2026, blending Gemini with Apple's Foundation Models under a roughly $1 billion collaboration. Apple is testing a redesigned, chat-like Siri in iOS 27/macOS 27, with a Dynamic Island interface and new features, ahead of a WWDC 2026 unveiling on June 8.

GateNews35m ago

SpaceX-Cursor $60B Deal Provides Fresh Evidence for SBF's Pardon Arguments

Abstract SpaceX announces a $60 billion Cursor partnership with an option to acquire, shaping SBF’s pardon bid as Alameda’s early Cursor stake would be worth about $3 billion today. The piece weighs SBF’s insolvency claims, parental campaigning, creditor objections, and the slim pardon odds. Summary SpaceX-Cursor deal fuels SBF pardon bid; Alameda’s $200k stake would be ~ $3B today (15,000x). SBF argues FTX insolvency; markets show low pardon odds; Trump unlikely to pardon.

GateNews43m ago

Chegg Stock Crashes 99% as AI Disrupts Edtech Market

Summary: Chegg soared during online-education demand, then AI tools disrupted its model, triggering massive layoffs and a collapse below $2, with broader AI-driven shifts hitting crypto miners and fintech firms. Abstract: This article examines Chegg's rise as a pandemic-era edtech darling and its ensuing decline amid the rapid adoption of generative AI, which provides quick answers and undercuts Chegg's value proposition. It documents 2025 layoffs and the stock's plunge toward delisting, and frames Chegg's experience within a broader AI disruption reshaping tech and crypto: Bitcoin miners pivot to AI operations, and AI-native strategies redefine competitiveness in fintech and beyond.

CryptoFrontier53m ago

OpenAI Releases Open-Source Privacy Filter Model for PII Detection and Redaction

Abstract: OpenAI's Privacy Filter is an open-source, locally executable model that detects and redacts PII in text. It supports large contexts, identifies many PII categories, and is intended for privacy-preserving workflows such as data preparation, indexing, logging, and moderation. OpenAI's Privacy Filter is a locally run, open-source model (128k-token context) that detects and redacts PII in text, covering contact, financial, and credential data for privacy workflows.

GateNews1h ago

OpenAI Plans to Deploy 30GW Computing Power by 2030

OpenAI aims for 30GW of computing by 2030 to meet rising AI demands, with 8GW completed of a 10GW 2025 target. The expansion signals a strategy to scale infrastructure for next-generation AI development and deployment. OpenAI intends to reach 30GW of computing power by 2030 to accommodate growing AI demands, having already completed 8GW of a 10GW target for 2025. The move reflects a strategic expansion of infrastructure to support next-generation AI development and deployment.

GateNews1h ago
Comment
0/400
No comments