ChatGPT Released NEW Major Changes That You Might’ve Missed (New ChatGPT Models)

Chat Gpt

ChatGPT has rolled out a set of updates that quietly change how you should use the platform. The biggest wins are practical: a new model selector that makes choosing the right brain simpler, access to newer model options (including cheaper ones through the API), and upgrades to both Sora for video creation and Codex for coding workflows. On top of that, ChatGPT is expanding integrations with everyday apps so it can pull from your tools instead of starting from scratch.

If you use ChatGPT for anything from content creation to research to agentic coding, these changes are worth understanding. The goal here is simple: help you set things up the right way so you get stronger results without wasting time or compute.

Table of Contents

The new ChatGPT model selector: instant, thinking, and pro

One of the most visible updates is the new model selector. Instead of hunting through a long list of options, ChatGPT now groups choices by what you are trying to do.

  • Instant for everyday chats
  • Thinking for complex questions
  • Pro (for research-heavy or technical work, like science and engineering)

This matters because “best model” often depends on the task type. A creative rewrite does not need the same level of reasoning as a technical plan or a deep analysis. The selector is designed to get you closer to the right mode faster.

Use the thinking level settings the smart way

Inside the “thinking” option, there are different thinking levels. The recommendation is straightforward:

  • If you are using thinking, default to a heavy setting.
  • If you are using pro, use extended thinking.
  • Avoid higher levels unless the task truly benefits from it.

Higher thinking effort is not automatically “better.” It is more deliberate. For tasks that need maximum capability (deep reasoning, complex problem solving, technical decision-making), it helps. For simpler tasks, it can slow you down without delivering proportional value.

Configure lets you toggle thinking on and off

There is also a configure area where you can choose between the latest models and control your “pro thinking effort” and whether it is turned on.

In practice, this gives you a workflow. If you want ChatGPT to feel snappy, configure it for instant. If you want to engage deeper reasoning, turn thinking on and select a higher effort setting. The key idea is to match the configuration to your goal for that specific session.

New ChatGPT models via API: GPT 5.4 mini and nano

Another upgrade is happening behind the scenes in the API model lineup. When you check the API documentation, you can see new model options such as:

  • GPT 5.4
  • GPT 5.0 mini
  • GPT 5.0 nano

The important detail is about availability. Some of these options are not necessarily shown the same way inside the main ChatGPT interface. However, if you are using the API directly, these model choices can be a big deal because they can be cheaper while still being very capable.

If your use case involves high volume (like summarization pipelines, tool-using agents, classification, or multi-step content generation), the cheaper models can improve cost efficiency dramatically.

Quick decision guide: when to pick which model

  • Use a smaller/cost-efficient model for repeatable tasks like extraction, rewriting, structured output, and routing.
  • Use a stronger model when the task requires nuanced reasoning or fewer retries.
  • Experiment with thresholds (for example, “send to nano unless confidence is low”).

As always, benchmark on your own prompts and output requirements. The right model is the one that reliably hits your quality bar at your target cost.

Sora updates: add references for consistent characters and settings

Sora has gained a feature that feels small but is a huge step forward: references. You can now upload images to define characters or settings you want to reuse in generated content.

Instead of hoping the model “sort of” keeps the same person across generations, you can anchor the style and identity. This is especially valuable for creators who work in series or build a recognizable visual world.

How reference uploads improve consistency

Here is what the workflow looks like conceptually:

  • Open the Sora creation flow and select references (under a beta section).
  • Click add new.
  • Upload an image of a character or a setting you want to reuse.
  • Name the reference (for example, “Rob headshot”).
  • Optionally adjust character attributes like style and configuration details.
  • Save the reference and reuse it in future creations.

The payoff is consistency. If you generate images or videos and the character keeps changing face, outfit, or proportions, that inconsistency becomes the bottleneck. References help remove that bottleneck so you can focus on the creative direction instead of re-prompting forever.

Better characters lead to better images and videos

Creators often run into a simple problem: AI output is impressive once, but inconsistent across iterations. References help you build continuity, which improves the quality of final assets and makes collaboration easier (designers and editors can count on stable identity and visual rules).

ChatGPT image upgrades: more styles, prompts, and libraries

Inside ChatGPT under images, you can now find a set of image styles and prompt suggestions, plus image libraries showing what has been created. The user-facing idea is clear: you can move faster from concept to finished output.

Instead of crafting every prompt from scratch, you can start with suggested prompts and steer from there. Libraries also matter because they let you reuse or compare results without losing your context.

When combined with model improvements and Sora’s reference support, these updates make visual workflows more repeatable. That is the real competitive advantage: less chaos, more control.

Use an optimized prompt workflow: MyPromptBuddy (and why it works)

One of the strongest practical takeaways is prompt optimization. Your prompt largely determines your output, especially for image and video generation where details like lighting, lens feel, hyper-realism cues, and scene composition can decide whether you get “solid” or “great.”

A tool mentioned in the workflow is MyPromptBuddy, a Chrome extension designed to generate an optimized prompt for different task types such as:

  • Standard prompts
  • Reasoning prompts
  • Deep research prompts
  • AI video prompts
  • AI image prompts

What the optimization process looks like

The workflow described goes like this:

  1. Start with your original request (example: generate an AI image of a Ferrari in a specific scene).
  2. Paste it into the extension.
  3. Define your ideal output (for example, “hyper realistic image”).
  4. Click optimize to generate a “super prompt.”
  5. Use that optimized prompt in ChatGPT to generate the image.

Why this changes results

Without prompt optimization, the output may be “good enough” but not truly match the realism and fidelity you want. With optimization, the prompt can include stronger constraints and more specific visual details, resulting in outputs that look more like the target (better model fidelity, better scene lighting, and fewer artifacts).

Even if you never use the same tool, the lesson is universal: treat prompts like a spec. Precision beats vibes.

Want to explore it? Start here: MyPromptBuddy.

Codex updates: more models, subagents, automations, and skills

Codex has received a cluster of improvements that make it feel more agentic and more “workflow-native.” The changes fall into four main areas: new models, subagents that work in parallel, automations, and the ability to install “skills” for extra capability.

1) More models inside Codex

New models are now available in Codex. More options typically mean better routing for different tasks, such as fast code edits versus deeper reasoning across a repository.

2) Subagents: delegate work in parallel

Subagents are a major conceptual change. Instead of Codex doing tasks in a strict sequence, you can spawn subagents that work concurrently.

For example, you can ask to:

  • Spawn a subagent to explore an area of the codebase.
  • Spawn another subagent to implement the requested changes.

This improves throughput. In real development, waiting for one subtask before starting the next is often wasteful. Parallelization reduces idle time and can help you reach results faster.

3) Automations: templates for repeatable engineering tasks

Codex now provides automations grouped into categories like:

  • Status reports
  • Release preparation
  • Incident triage
  • Code quality
  • Repo maintenance
  • Growth and exploration

Instead of designing an automation from scratch each time, you can pick a task, choose the work tree or project scope, and create an automation from a template (or skip templates if you prefer).

This is huge for teams because it standardizes engineering workflows and reduces “tribal knowledge.”

4) Skills: upload capabilities for specific tools and workflows

Codex can now accept skills that extend its capability for your environment. The idea is to install the tools you rely on so Codex can interact with them more effectively.

Examples mentioned include:

  • Notion workflows
  • Playwright for browser automation
  • Screenshot capture

Skills turn Codex from “generic coding assistant” into a tool that understands your stack.

ChatGPT app integrations: connect Gmail, Slack, and more for deep research

Beyond models, ChatGPT is expanding the places it can operate. Integrations with apps mean ChatGPT can use your existing context rather than asking you to paste everything manually.

In the Apps section, you can browse categories such as Lifestyle and Productivity. These categories are meant to make common tools easy to discover and connect.

Connect apps as sources

One workflow described is adding “sources” for tasks. For example, you can connect:

  • Slack
  • Gmail
  • Other connected tools

Then, inside a task like deep research, ChatGPT can follow specific websites or use connected apps to retrieve information.

Deep research can build reports using your connected tools

An example provided: deep research connected with real estate and travel tools. The system ran multiple searches, cited sources, and created a report for a decision like where to buy a house in Montana. It also mapped travel timing for flights and hotels using Expedia, including estimates such as round-trip cost and five-star hotel options.

The larger takeaway is that “deep research” is shifting from “read the web” to “query your tool ecosystem.” That means faster workflows and fewer manual steps.

If you want practical value, start by connecting the one or two apps you actually use daily. Then build from there. Integrations are only useful if they reduce effort in your real routines.

What to do next: a simple upgrade checklist

Here is a quick, actionable checklist based on these updates.

  1. Update your model habits
    • Use Instant for routine chat.
    • Use Thinking for tougher questions.
    • Use Pro for research, science, engineering, and deep reasoning.
  2. Match thinking effort to the task
    • Heavy for general thinking tasks.
    • Extended when you truly need maximum reasoning.
  3. If you use the API, check the model lineup
    • Look for cheaper options for high-volume or structured work.
    • Benchmark quality and cost together.
  4. For images and video, anchor your identity
    • Use Sora references to keep characters consistent across iterations.
    • Use libraries and style presets for faster iteration.
  5. Upgrade your prompts
    • Use prompt optimization to get closer to hyper-realistic or specific visual outcomes.
    • Write prompts like specs, not requests.
  6. For coding workflows, try Codex subagents and automations
    • Spawn subagents for parallel exploration and implementation.
    • Use automations for repeatable tasks.
    • Install skills for your toolchain.
  7. Connect apps if you want real-world research speed
    • Start with Gmail and Slack or your closest equivalent.
    • Use deep research to generate reports based on connected sources.

FAQ

What is the new ChatGPT model selector, and why does it matter?

It groups models into practical categories: Instant for everyday chats, Thinking for complex questions, and Pro for research-heavy technical tasks. It helps you pick the right level of reasoning faster and avoid unnecessary compute.

Should I always use “extended” thinking in ChatGPT Pro?

No. “Extended” thinking is best when the task benefits from maximum reasoning. For simpler tasks, it can slow you down without improving results proportionally.

How do GPT 5.4 mini and nano relate to ChatGPT?

Those model options are visible in the API documentation and are positioned as cheaper models for agentic or high-volume tasks. Some are used implicitly for “Instant” experiences, but the API is where you can intentionally select them.

What are Sora references?

References let you upload images of characters or settings and reuse them in new video or image generations. This improves consistency, especially for series-style content.

Why do optimized prompts produce better image results?

Image generation is sensitive to prompt detail. Optimized prompts can add clearer constraints about realism, lighting, composition, and style, reducing ambiguity and improving fidelity.

What’s new in Codex with subagents and automations?

Subagents allow parallel work, automations provide template-driven workflows like triage and release prep, and skills enable Codex to better interact with your specific tools such as Notion or Playwright.

Leave a Reply

Your email address will not be published. Required fields are marked *

Most Read

Subscribe To Our Magazine

Download Our Magazine