← 2026-03-04

Daily Edition

2026-03-05

2026-03-06 →

X / Twitter

17
John Carmack
John Carmack @ID_AA_Carmack
Trista and I saw NIN last night with friends, and it was a great show, as always.

I inevitably wind up thinking about the gap between the live event experience and reproductions on video or in VR. The gap is substantial, with the “shock and awe” of high powered, extreme dynamic range stadium effects, but there are still levers to pull.

A capability of the Quest VR hardware that would be fun to exploit is per-frame varying of the display persistence. The low-persistence display is only flashed “on” for about a tenth of each frame, so it doesn’t look blurred as you move your head around. This is a tradeoff between screen brightness and blur, which is just statically picked by Meta today, but it could be done dynamically on a per-frame basis (coordination would be tricky!).

Going to full persistence briefly would give an instant 10x brightness increase, which would be quite impactful on eyes adapted to the 10% brightness. You wouldn’t want to keep it there for long, because the blurring would be obnoxious and eyes would adapt to it, but it would be perfect for the epilepsy-inducing concert light flashes.

The clean way to do it would be with a 16 bit high dynamic range linear buffer that gets factored into an sRGB image for the display and a backlight flight time, but a hacky extra parameter latched with the frame submit would be the easiest thing to experiment with. Make it work, then make it clean!

A tech demo would be fun, but Beat Saber has a large enough user base that stepping up “concert lighting effects” might even be worth it from a net user value standpoint.

Trista: NIN - Friggin 🔥

ID_AA_Carmack
ID_AA_Carmack @ID_AA_Carmack
Trista and I saw NIN last night with friends, and it was a great show, as always.

I inevitably wind up thinking about the gap between the live event experience and reproductions on video or in VR. The gap is substantial, with the “shock and awe” of high powered, extreme dynamic range stadium effects, but there are still levers to pull.

A capability of the Quest VR hardware that would be fun to exploit is per-frame varying of the display persistence. The low-persistence display is only flashed “on” for about a tenth of each frame, so it doesn’t look blurred as you move your head around. This is a tradeoff between screen brightness and blur, which is just statically picked by Meta today, but it could be done dynamically on a per-frame basis (coordination would be tricky!).

Going to full persistence briefly would give an instant 10x brightness increase, which would be quite impactful on eyes adapted to the 10% brightness. You wouldn’t want to keep it there for long, because the blurring would be obnoxious and eyes would adapt to it, but it would be perfect for the epilepsy-inducing concert light flashes.

The clean way to do it would be with a 16 bit high dynamic range linear buffer that gets factored into an sRGB image for the display and a backlight flight time, but a hacky extra parameter latched with the frame submit would be the easiest thing to experiment with. Make it work, then make it clean!

A tech demo would be fun, but Beat Saber has a large enough user base that stepping up “concert lighting effects” might even be worth it from a net user value standpoint.

Trista: NIN - Friggin 🔥

Aaron Levie
Aaron Levie @levie
Agents will be the biggest users of software. They’ll often need their own computers, identities, file systems, and tools to do their work. As a result, software will increasingly become API-first to be as useful to agents as they are to people. This is a huge opportunity.

Box: Agents need files to keep track of their work, they leverage files as context about the tasks they’re doing, and use them to share back and forth with their human counterparts. @levie spoke with @CNBC about and the importance of agents having their own filesystems.

Ryo Lu
Ryo Lu @ryolu_
Make agents work while you think, while you play, while you sleep.

This is Cursor Automations.

Cursor: We're introducing Cursor Automations to build always-on agents.

Kevin Weil 🇺🇸
Kevin Weil 🇺🇸 @kevinweil
Retweeted
Derya Unutmaz, MD Derya Unutmaz, MD
I’ve had early access to GPT-5.4 Pro. Without any reservation, I can say it is the most intelligent AI model to date, even significantly surpassing GPT-5.2 Pro at several levels! I’ve been using it non-stop past several days and am super excited about another major jump in AI!
I will share specific examples, but overall GPT-5.4 Pro demonstrates relatively higher creativity, insight, and abstract intelligence. It tends to ask “why,” “what if,” “can I,” and “why it matters” type questions more frequently than the 5.2 Pro model.
It also appears to generalize more effectively and comes across as more AGI-like in its reasoning, and even displays human-like intuition! Especially biomedical science-based responses are unifying large data sets and simply amazing!
Sam Altman
Sam Altman @sama
Retweeted
Noam Brown Noam Brown
GPT-5.4 is a big step up in computer use and economically valuable tasks (e.g., GDPval). We see no wall, and expect AI capabilities to continue to increase dramatically this year.
OpenAI: GPT-5.4 Thinking and GPT-5.4 Pro are rolling out now in ChatGPT.
GPT-5.4 is also now available in the API and Codex.
GPT-5.4 brings our advances in reasoning, coding, and agentic workflows into one frontier model.
Kevin Weil 🇺🇸
Kevin Weil 🇺🇸 @kevinweil
💥 GPT 5.4 is launching today! It's our best model ever, and it's also the most capable scientific model we've ever released.

GPT 5.4 Pro in particular is 🤯 based on early testing with scientists and mathematicians.
Sam Altman
Sam Altman @sama
GPT-5.4 is launching, available now in the API and Codex and rolling out over the course of the day in ChatGPT.

It's much better at knowledge work and web search, and it has native computer use capabilities.

You can steer it mid-response, and it supports 1m tokens of context.
Sam Altman
Sam Altman @sama
Codex app on Windows!

Andrew Ambrosino: The Codex app is now live on Windows.

The app runs both natively and in WSL, with integrated terminals for PowerShell, Command Prompt, Git Bash, or WSL.

We also built the first Windows-native agent sandbox — using OS-level controls to block filesystem writes outside your

Sam Altman
Sam Altman @sama
Forgot to mention /fast!

I think people will like this.

Ahmed: Today we are introducing GPT-5.4 in codex. It's more token efficient and better at tool calling, computer use, and frontend development.

We are also introducing /fast to get a faster version of Codex.

Enjoy ❤️

Kevin Weil 🇺🇸
Kevin Weil 🇺🇸 @kevinweil
Retweeted
Epoch AI Epoch AI
GPT-5.4 set a new record on FrontierMath, our benchmark of extremely challenging math problems! We had pre-release access to evaluate the model. On Tiers 1–3, GPT-5.4 Pro scored 50%. On Tier 4 it scored 38%.
See thread for commentary and additional experiments.
Sam Altman
Sam Altman @sama
We will be able to fix these three things!

Matt Shumer: I've been testing GPT-5.4 for the last week.

In short, it is the best model in the world, by far.

It's so good that it's the first model that makes the “which model should I use?” conversation feel almost over.

The biggest surprise: I barely use Pro anymore!

If you know me,

Fei-Fei Li
Fei-Fei Li @drfeifei
Retweeted
World Labs World Labs
70 hackers joined us in SF for the first-ever World Labs Hackathon.
In just 3.5 hours, 32 teams used Marble for projects ranging from robotics sims and agents to AR/VR interfaces, games, art experiences, and real estate tools.
Check out what they built ↓
drfeifei
drfeifei @drfeifei
Retweeted
World Labs World Labs
70 hackers joined us in SF for the first-ever World Labs Hackathon.
In just 3.5 hours, 32 teams used Marble for projects ranging from robotics sims and agents to AR/VR interfaces, games, art experiences, and real estate tools.
Check out what they built ↓
Aaron Levie
Aaron Levie @levie
Model progress continues unabated, with GPT-5.4 showing significant improvements in critical knowledge worker tasks. In our Box AI tests, we saw a 6 point jump in agentic document processing, which is upstream from most automation workflows. GPT-5.4 is now available on Box.

Box: We tested @OpenAI's new GPT-5.4 model, and it showed a 78% overall extraction accuracy - up 6 points from GPT-5.2.

Our evaluation tested the model across real industries and document workflows:
→ Clinical data: +5 pts (81% → 86%)
→ Legal agreements: +3 pts (82% → 85%)


Andrej Karpathy
Andrej Karpathy @karpathy
There was a nice time where researchers talked about various ideas quite openly on twitter. (before they disappeared into the gold mines :)).

My guess is that you can get quite far even in the current paradigm by introducing a number of memory ops as "tools" and throwing them into the mix in RL. E.g. current compaction and memory implementations are crappy, first, early examples that were somewhat bolted on, but both can be fairly easily generalized and made part of the optimization as just another tool during RL.

That said neither of these is fully satisfying because clearly people are capable of some weight-based updates (my personal suspicion - mostly during sleep). So there should be even more room for more exotic approaches for long-term memory that do change the weights, but exactly - the details are not obvious. This is a lot more exciting, but also more into the realm of research outside of the established prod stack.

Awni Hannun: I've been thinking a bit about continual learning recently, especially as it relates to long-running agents (and running a few toy experiments with MLX).

The status quo of prompt compaction coupled with recursive sub-agents is actually remarkably effective. Seems like we can go
swyx
swyx @swyx
Retweeted
William Shen William Shen
Excited to introduce Uni-1, our new multimodal model that *unifies* understanding and generation.
TLDR: a team of ~15 researchers is going pound-for-pound with nano banana and gpt image 🧵
Jiaming Song: Excited to introduce Uni-1, our new *unified* multimodal model that does both understanding and generation: https://lumalabs.ai/uni-1
TLDR: I think Uni-1 @LumaLabsAI is > GPT Image 1.5 in many cases, and toe-to-toe with Nano Banana Pro/2. (showcase below)

YouTube

2
No Priors: AI, Machine Learning, Tech, & Startups

The Biggest Misconception About AI Debt

Watch Video
Latent Space

Why Every Agent Needs a Box — Aaron Levie, Box

Watch Video
View Transcript (0 words)