Language:English VersionChinese Version

Introduction

The following content has been compiled by VIPSTAR, based on public information from X / social media. It is intended for reading and research purposes only.

Key Points

  • It was around 2 AM, and I was just about to go to bed.
  • Suddenly, GPT-5.4 was released.

Note

For any sections involving rules, benefits, or judgments, please refer to the original statements and the latest official information from Digital Life Katsik.

Editor’s Comment

The article titled “X Import: Digital Life Katsik – GPT-5.4 Midnight Release, the Perfectly Suited Model for OpenClaw Finally Arrives” comes from the X social platform, authored by Digital Life Katsik. From a content completeness perspective, the original text provides a high density of key information, particularly in terms of core conclusions and actionable recommendations. It was around 2 AM when I was just about to go to bed, and then GPT-5.4 was suddenly released. I got so excited that I couldn’t sleep at all. Seriously, this isn’t me being overly dramatic; it’s rare for me to use such an expression of excitement. The reason is that I’ve been waiting for the official release of GPT-5.3 or GPT-5.4 to serve as my preferred model for OpenClaw. The reasoning is quite straightforward: in the modern world, which has fundamentally been built on code over the past thirty years, everything we see about computers and the internet… For readers, its most direct value isn’t just “learning a new perspective,” but being able to quickly understand the conditions, boundaries, and potential costs behind that perspective. If we break down this content into verifiable statements, it includes at least the following aspects: It was around 2 AM when I was just about to go to bed; then GPT-5.4 was suddenly released. Among these statements, the conclusions are often the easiest to spread, but what truly determines their practicality is whether the underlying assumptions hold true, if the sample size is sufficient, and if the time frame matches. We recommend that readers, when referencing such information, prioritize verifying data sources, publication times, and any differences in platform environments to avoid mistaking “contextual experience” for “universal rules.” From an industry impact perspective, this type of content typically exerts a short-term guiding influence on product strategies, operational rhythms, and resource allocation, especially in themes related to AI, development tools, growth, and monetization. From the editor’s viewpoint, we are more concerned with whether it can withstand subsequent factual tests: 1) Can the results be replicated? 2) Is the method transferable? 3) Are the costs sustainable? The source is x.com, and we suggest readers consider this as one of several inputs for decision-making rather than a sole basis. Finally, here’s a practical suggestion: If you plan to act on this information, start with a small-scale test and gradually increase your investment based on feedback. If the original content involves profits, policies, compliance, or platform rules, refer to the latest official announcements and have a rollback plan in place. The purpose of re-posting is to enhance the efficiency of information circulation, but the true value of content is realized through secondary judgment and localized practice. Based on this principle, our accompanying editorial comments will continuously emphasize verifiability, boundary awareness, and risk control to help you transform “information seen” into “actionable knowledge.”

It was around 2 AM, and I was just about to go to bed.

Then, GPT-5.4 was suddenly released.

![Image](https://novvista.com/wp-content/uploads/2026/03/HCshf1IaUAEJFZj-1.png)
*Figure 1: Accompanying Image*

I got so excited that I couldn’t sleep at all.

Really, this isn’t me being overly dramatic. I rarely use phrases like “too excited to sleep.”

The reason is that I’ve been waiting for the official release of GPT-5.3 or GPT-5.4 to serve as the primary model for my OpenClaw project.

The reasoning is quite straightforward: in the modern world, which has fundamentally evolved over the past thirty years, everything we see about computers and the internet is built on a foundation of code.

So you can understand that coding ability often represents one strong leg of an Agent’s capabilities.

In my view, an excellent base model for an Agent generally needs to excel in three areas:

Coding ability, world knowledge, and multimodal understanding.

When all three are state-of-the-art (SOTA), you’re almost certainly dealing with the best Agent model. Of course, another crucial factor is cost.

In the past, Claude Opus 4.6 was practically synonymous with Agent models because it excelled in coding and world knowledge. Although its multimodal capabilities weren’t as strong as Seed 2.0 or Gemini 3.1 Pro, they were sufficient for many scenarios since current Agents don’t interact much with the physical world—that’s more in the realm of embodied intelligence.

GPT-5.3-Codex, which I used to love, had exceptional coding ability and was incredibly precise when executing tasks.

However, the biggest issue was that it was a specialized programming model. Its world knowledge was terrible, even worse than GPT-5.2. So, OpenAI had no choice but to release it with the Codex suffix to compete with Claude.

So you can see…

In terms of planning capabilities, it is completely outmatched by Claude Opus 4.6. However, the biggest issue is actually due to its lack of world knowledge, which causes a lot of problems.

It speaks in gibberish; the things it says are really hard for me to understand. I’m not a programmer by training, and reading what it writes is genuinely very challenging.

For example, I asked it to review one of my AI trend website projects, mainly to check my documentation standards and the entire codebase.

Then, the document this guy wrote… I mean…

![Image 2](https://novvista.com/wp-content/uploads/2026/03/HCshgShaUAMOpu2-1.png)

Compare it with what Claude Opus 4.6 wrote.

![Image 3](https://novvista.com/wp-content/uploads/2026/03/HCshgwhaUAEvE2T-1.png)

The difference should be obvious…

Because this thing doesn’t speak human language and lacks world knowledge, it’s fine to use it within Codex, but if you integrate it into your OpenClaw as the default model, you’ll realize what a disaster it is. This guy sounds so inhuman that I feel like punching him when he talks.

So, I tried it once and immediately abandoned it. In my OpenClaw, I still use Claude Opus 4.6 and Sonnet 4.6 for scene calls.

Why am I looking forward to GPT-5.4?

Because, while Claude is great in every way, it’s just too expensive!!!

It’s really, really expensive!!!!

And because of the clueless Anthropic, they’ve made OpenCla… (the text seems to be cut off here)

W went crazy, so the quota I subscribed to for Claude’s Max Plan can only be used on Claude Code and not at all for OpenClaw. If you want to use it on OpenClaw, you have to directly input the API Key.

However, everyone knows how expensive Claude’s API is; it’s something our broke team simply can’t afford. It might work for small-scale usage, but large-scale use would bankrupt the company.

There used to be another way: using a reverse proxy to route Claude’s quota from Google’s Antigravity through a plugin and then pass it to OpenClaw.

![Image 4](https://novvista.com/wp-content/uploads/2026/03/HCshhO2aUAMa1PK-1.png)

But later, Google started suspending accounts in large numbers, making this method unusable.

During the New Year, my Google account was also suspended, forcing me to use AI to write a heartfelt and tearful email to Google.

![Image 5](https://pbs.twimg.com/media/HCshhooaUAIo-_z.jpg)

I said I was wrong and that it would never happen again.

Only then did Google unblock my account, but the reverse proxy method is definitely out of the question now.

OpenAI, on the other hand, took a different approach. When Claude started aggressively suspending OpenCode accounts, OpenAI stepped up and declared that they wouldn’t do any suspensions, encouraging everyone to use their services freely.

![Image 6](https://novvista.com/wp-content/uploads/2026/03/HCshiHJaUAI5JZ3-1)

Figure 6

This is the only one among the Big Three that supports using third-party tools to call on Codex’s quota.

The same goes for OpenClaw. It’s one of the few top-tier models that can be accessed directly through a login, while others require an API.

Figure 7

Really, OpenAI has been incredibly generous this time around.

They’ve also been ramping up the quota for Codex.

Figure 8

So, while Claude is great to use in OpenClaw, it can only be used via API and that gets quite expensive since you can’t use the subscription quota.

OpenAI’s models, on the other hand, do allow for the use of subscription quotas, but GPT-5.2 doesn’t handle code well, and GPT-5.3-Codex doesn’t understand human language properly.

You see how awkward that is?

But now, here comes GPT-5.4!

Finally, it fills in the gaps!

It matches GPT-5.3-Codex in coding ability, surpasses GPT-5.2 in world knowledge, and you can use your subscription quota with it. For just $20, you get an incredibly smooth experience.

Tell me, if this isn’t the perfect model for OpenClaw, who is? Hmm?

Starting today, everyone using OpenClaw should set their default model to…

Switch to GPT-5.4, really, trust me.

Back to GPT-5.4, as usual, let’s start with the benchmarks.

Image 9

It’s really satisfying.

Let’s look at the most crucial ones first.

GDPval: 83.0%

This measures AI performance in real-world tasks, covering knowledge work across 44 professions including finance and law.

GPT-5.4 Thinking scored 83.0%, Claude Opus 4.6 scored 78.0%, and GPT-5.3 Codex scored 70.9%.

In real business scenarios, GPT-5.4 isn’t just good at writing code; it can also discuss business, finance, law, and various professional topics with you.

And it does so in human language, not gibberish.

SWE-Bench Pro: 57.7%

This measures AI’s ability to solve real software engineering problems, not just in Python but across four programming languages.

GPT-5.4 Thinking scored 57.7%, while GPT-5.3 Codex scored 56.8%.

They are basically on par.

This is exactly what I wanted to see.

GPT-5.4 maintains the coding capabilities of GPT-5.3 Codex and also improves its world knowledge.

The same goes for OSWorld-Verified, which scored 75.0%. This measures AI’s ability to operate a computer, such as using a mouse, keyboard, switching between applications, and completing various tasks.

GPT-5.4 Thinking scored 75.0%, surpassing Claude Opus 4.6’s 72.7% and maintaining parity with GPT-5.3-Codex.

Moreover, the speed at which GPT-5.4 operates a computer is incredibly fast.

Figure 10

Watch this video without acceleration to get a clearer understanding.

ToolAthlon: 54.6%

This measures the AI’s ability to use tools, which is one of the core indicators of an agent’s capabilities.

GPT-5.4 Thinking scored 54.6%, while Claude Sonnet 4.6 scored 44.8%.

There’s nearly a 10-point difference.

As for academic knowledge, it can’t compare to GPT-5.3-codex because OpenAI is aware of this and didn’t even run the test at that time.

Figure 11

In simpler terms, it means:

GPT-5.4 = GPT-5.3 Codex’s coding ability + world knowledge surpassing GPT-5.2 + enhanced tool usage capability + extremely affordable codex quota.

When you combine these four aspects, you get a perfect OpenClaw base model.

Additionally, there are several excellent feature updates:

  1. A 1 million token context window.

This is a significant upgrade for GPT-5.4.

Previously, GPT-5.3 had a context window of 400,000 tokens, and GPT-5.4 has more than doubled this to 1 million tokens.

This is crucial for agents because they need to maintain an understanding of the entire task context while executing tasks. For example,

If the context window is not large enough, the Agent may start to forget things as it works, and won’t remember what was mentioned earlier.

One million tokens are basically sufficient to handle most Agent tasks.

Of course, OpenAI isn’t foolish; they state that after 270,000 tokens, your quota will be counted as double.

![Figure 12](https://novvista.com/wp-content/uploads/2026/03/HCshlYjbsAAvuoe-1.png)

However, because the Codex quota is so generous, even doubling it is still manageable.

1. Native computing capabilities.

This is another major selling point of GPT-5.4.

OpenAI claims that GPT-5.4 is their first mainline model with built-in native computing capabilities.

It excels at writing code to manipulate computers using libraries like Playwright and can also issue mouse and keyboard commands based on screenshots.

In other words, it combines coding and visual processing. I feel that once this feature is integrated, it will truly allow for direct visual control of most software on your computer, which is really exciting.

Based on this, they have released a new skill called `playwright-interactive`.

This allows Codex to debug Web and Electron applications using both code and visual methods simultaneously.

![Figure 13](https://novvista.com/wp-content/uploads/2026/03/HCshlx1aUAAweSa-1.jpg)

You can find the link [here](#), where you can install it yourself.

https://github.com/openai/skills/tree/main/skills/.curated/playwright-interactive

  1. Tool search support has been added.

In the past, when a model was provided with tools, all tool definitions were included in the prompt beforehand.

For systems with a large number of tools, this could add thousands or even tens of thousands of tokens to each request, which is often meaningless. This unnecessarily increases costs, slows down responses, and fills the context with information that the model may never use.

To address this, they have now added tool search support. GPT-5.4 no longer receives complete tool definitions directly but instead gets a lightweight list of available tools along with a tool search function.

When the model needs to use a specific tool, it can look up the tool’s definition and append it to the conversation at that moment.

This approach is similar to the progressive disclosure method used in Skills. The goal is simple: to optimize context engineering.

After testing internally, OpenAI found that the tool search configuration reduces overall token usage by 47% while maintaining the same accuracy. This is a significant improvement.

Image 14

GPT-5.4 Thinking is roughly as described above.

This time, they also released GPT-5.4 Pro, which I won’t go into detail about. Essentially, everything is more powerful, but for most people, it’s too expensive and not very practical, costing $200.

Sure, here is the translation:

Only Pro members can use it.

It’s still worth mentioning the overall API pricing, even though most users will likely opt for the subscription plan.

Image 15 accompanying the text

Compared to GPT-5.2, the price has increased, but it is still significantly cheaper than Claude Opus 4.6. The price for Claude Opus 4.6 is $5/$25 per million tokens (input/output), while GPT-5.4 costs only half as much.

ChatGPT is now live.

Image 16 accompanying the text

Codex is also supported, and I had a brief experience with it myself.

Image 17 accompanying the text

The first thing that struck me was the refreshingly human-like responses…

For example, when I asked it to download a video from the OpenAI website, its response was: “This is the most annoying task,” and “It saves both of us from Cloudflare’s lifespan drain”…

I hope this translation meets your needs!

Image 18

And this one.

Really, I can actually understand the output from Codex…

Image 19

The design of the front end has made some nice progress, but it still doesn’t match up to Opus 4.6 and Gemini.

Image 20

I did a rough test of the writing, and it still has that peculiar tendency to use parallel sentences in a somewhat bizarre way.

Quite strange.

Then, it’s a bit disappointing that I waited until around 6 AM, but OpenClaw’s current method of logging in with Codex still doesn’t support GPT-5.4.

Image 21

This means I still haven’t had the chance to test GPT-5.4 on crayfish.

Sure, here is the translation:

Effect.

However, I think today’s update will support it.

This is because many users in the community have been urging for it, and early adopters generally report very positive results.

I’m eagerly waiting for the support; I can hardly wait anymore.

Another exciting night ahead.

If you’re using OpenClaw, remember to switch the default model to GPT-5.4 once it’s supported.

If you haven’t tried OpenClaw yet, now is a great time to start.

After all, with the divine GPT-5.4 model, the experience will only be better.

The year 2026 is truly insane.

Source
Author: Digital Life Khazix
Publish Date: March 6, 2026, 11:10 AM
Source: Original Post Link

By

Leave a Reply

Your email address will not be published. Required fields are marked *