×
OpenAI unveils advanced visual AI models with superior image processing capabilities
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

OpenAI‘s latest AI models, o3 and o4-mini, represent a significant leap forward in capabilities for ChatGPT, offering enhanced multimodal understanding and greater autonomy when solving complex problems. These new “brains” for ChatGPT arrive as the company benefits from unprecedented computing resources, signaling a pivotal moment in AI development where processing power constraints no longer limit innovation potential.

The big picture: OpenAI has released two advanced AI models, o3 and o4-mini, exclusively for ChatGPT Plus, Pro, and Team users, while announcing the retirement of GPT-4 by month’s end.

  • These models feature significantly improved ability to interpret poor-quality images, including blurry photos and handwritten notes, enhancing their multimodal capabilities.
  • The update comes days after ChatGPT received major memory enhancements designed to make interactions more personalized and context-aware.

Key capabilities: The new models can independently utilize all of ChatGPT’s integrated tools without requiring explicit user instructions.

  • For the first time, the AI can autonomously decide when to deploy the browser, Python code interpreter, image generation, or image analysis tools based on the task at hand.
  • Users can now interact with images in real-time, performing operations like rotating or zooming during problem-solving sessions.

Developer focus: OpenAI is launching Codex CLI, a new open-source coding agent that runs locally in terminal windows and works with these new models.

  • The company is offering $1 million in API credits distributed in $25,000 increments to encourage developers to build with these tools.
  • A more advanced o3-pro model is scheduled for release in the coming weeks, while Pro plan users can continue using the existing o1-pro model.

Why this matters: OpenAI CEO Sam Altman recently revealed the company is no longer “compute constrained,” marking a significant shift in development capabilities.

  • This newfound processing capacity enables faster development cycles and the creation of models capable of handling increasingly complex tasks.
  • The strategic focus on autonomy and tool integration points toward OpenAI’s vision of AI systems that function less as simple assistants and more as collaborative partners for creative and analytical work.
OpenAI just launched its smartest AI yet that can think with images — here's how to try it

Recent News

Scaling generative AI 4 ways from experiments to production

Organizations face significant hurdles when moving generative AI initiatives from experimentation to production-ready systems, with most falling short of deployment goals despite executive interest.

Google expands Gemini AI with 2 new plans, leak reveals

Google prepares to introduce multiple subscription tiers for Gemini, addressing the gap between its free and premium AI offerings.

AI discovers potential Alzheimer’s cause and treatment

AI identifies PHGDH gene as a direct cause of Alzheimer's disease beyond its role as a biomarker, offering a new understanding of spontaneous cases and potential treatment pathways.