Image created with gemini-3.1-flash-image-preview with claude-sonnet-4-5. Image prompt: Wide shot of a drab Chinese municipal waiting room with worn plastic chairs and peeling walls, one person sitting alone, fluorescent light, through the window a chestnut horse stands in a concrete courtyard, faded Chinese administrative poster on wall reading ‘intelligent services’, overcast daylight, muted grays and greens, observational realism, large white text overlay reading ‘OpenAI’ positioned like Chinese cinema poster title, documentary stillness, decelerated time, human-scale composition

Excited to work with Peter Steinberger to build the future of agents for everyone and to continue to improve Codex in leaps and bounds. We are committed to OSS, continuing to make OpenClaw flourish and bringing agents to life in a way that is fun, safe and highly productive.”” https://x.com/thsottiaux/status/2023147973421785386

There have been fair questions on whether LLM contributions to STEM are overhyped, but I’ve spoken with physicists about this result and they’ve told me it is a truly significant research contribution, roughly at the level of a solid journal paper, and GPT-5.2 played a key role.”” https://x.com/polynoamial/status/2022413904757035167

OpenAI plans AI device lineup, including speaker and smart glasses”” https://x.com/Crypto_Briefing/status/2024890816167116956

OpenAI’s acquisition of OpenClaw signals the beginning of the end of the ChatGPT era | VentureBeat https://venturebeat.com/technology/openais-acquisition-of-openclaw-signals-the-beginning-of-the-end-of-the

GPT 5.2 derived a new result in theoretical physics. For decades it’s been assumed that certain gluon amplitudes (“”single minus””) were zero, and that the maximally helicity violating amplitudes had two gluons of one helicity and n-2 of the other. It turns out that isn’t”” https://x.com/kevinweil/status/2022388305434939693

GPT-5.2 derived a novel result in theoretical physics, showing that a type of particle interaction many physicists expected would not occur can in fact arise under specific conditions. There is great promise in the potential of AI to benefit people by accelerating science.”” https://x.com/gdb/status/2022394113971360145

GPT-5.2 derives a new result in theoretical physics | OpenAI https://openai.com/index/new-result-theoretical-physics/

I spent last night with Andrew Strominger and Alex Lupsasca, two of the top physicists in the world They just released a paper, co-authored with OpenAi, that seems to me like ASI Andrew, who helped develop string theory, told me that a year ago, his view was that he didn’t know”” https://x.com/patrick_oshag/status/2022395157648195801

More on the gluon scattering/GPT 5.2 paper from @ALupsasca below 👇 If you’re in the Boston area on Tuesday, go see his lecture at Harvard!”” https://x.com/kevinweil/status/2023422106411974935

We’re committing $7.5M to @AISecurityInst’s Alignment Project to fund independent research on mitigations for safety and security risks from misaligned AI.”” https://x.com/OpenAINewsroom/status/2024546609485533442

Introducing EVMbench | OpenAI https://openai.com/index/introducing-evmbench/

Introducing EVMbench–a new benchmark that measures how well AI agents can detect, exploit, and patch high-severity smart contract vulnerabilities.”” https://x.com/OpenAI/status/2024193883748651102

OpenAI Funding on Track to Top $100 Billion in Latest Round – Bloomberg https://www.bloomberg.com/news/articles/2026-02-19/openai-funding-on-track-to-top-100-billion-with-latest-round

Peter Steinberger is joining OpenAI to drive the next generation of personal agents. He is a genius with a lot of amazing ideas about the future of very smart agents interacting with each other to do very useful things for people. We expect this will quickly become core to our”” https://x.com/sama/status/2023150230905159801

Wowww…….big news Charles Porch who has been VP of global partnerships for IG is going to OpenAI.”” https://x.com/yashar/status/2024187504682029171

Introducing Lockdown Mode and Elevated Risk labels in ChatGPT | OpenAI https://openai.com/index/introducing-lockdown-mode-and-elevated-risk-labels-in-chatgpt/

🦞”” https://x.com/sama/status/2023463428892094655

After initially being hyped about the speed, I have to say that 5.3-codex-spark, even on xhigh, is actually quite a bit dumber than 5.3-codex, to the point that I’m back to using the latter most of the time.”” https://x.com/giffmana/status/2023341811851473053

agents are up and productive, time for bed”” https://x.com/gdb/status/2023342301821734937

codex is so good at the toil — fixing merge conflicts, getting CI to green, rewriting between languages — it raises the ambition of what i even consider building”” https://x.com/gdb/status/2023135825970749637

codex momentum is strong, and many people are feeling just how big of a leap 5.3 is. if your organization hasn’t tried codex yet, it’s worth revisiting.”” https://x.com/gdb/status/2023299087974777061

Codex weekly users have more than tripled since the beginning of the year!”” https://x.com/sama/status/2023233085509410833

codex’s shell-fu is incredible to behold and learn from”” https://x.com/gdb/status/2022823856889827711

I am increasingly asked during candidate interviews how much dedicated inference compute they will have to build with Codex. Pairing this with usage per user growing significantly faster than the number of users, it’s pretty clear that compute will be something that is scarce.”” https://x.com/thsottiaux/status/2024635825997459841

I’m glad that 5.3 Codex has started making good use of sub-agents. With 5.2 Codex, I often saw it not using them much even when the feature was enabled. For reference: I wouldn’t recommend this for Plus users, but for Pro users, you can increase the maximum number of sub-agents”” https://x.com/Hangsiin/status/2023297599764402627

I’ve had 3 or more agents running in parallel with Codex for 2+ hours. I’ve used 8% of my 5-hour window. 2% of my weekly. I am literally trying to hit the limits and still can’t.”” https://x.com/theo/status/2023718038198251904

measuring agentic security capabilities with smart contracts:”” https://x.com/gdb/status/2024200501055963593

We have a special thing launching to Codex users on the Pro plan later today. It sparks joy for me. I think you are going to love it…”” https://x.com/sama/status/2021984777470193767

On evaluating multi-step scientific tool use in LLM agents. SciAgentGym provides an interactive environment with 1,780 specialized tools across 4 scientific disciplines. The core finding: even advanced models like GPT-5 see success rates drop sharply from 60.6% to 30.9% as”” https://x.com/dair_ai/status/2023404773031166320

One thing I feel not enough people know is that the Codex agent is open source. It also exposes an app-server interface that lets you integrate Codex into your application including sign-in with ChatGPT. It’s the same server that powers Codex in VSCode, Jetbrains and Xcode”” https://x.com/dkundel/status/2024233673764257879?s=20

Introducing Lockdown Mode for ChatGPT. Lockdown mode is an advanced, optional security setting for higher-risk users, businesses, and enterprises. Lockdown Mode disables certain tools and capabilities in ChatGPT that an adversary could attempt to exploit to exfiltrate sensitive”” https://x.com/cryps1s/status/2023441322838028362

Wow, Codex is some sort of a miracle… (yes, I’ve tried Claude Code before that)”” https://x.com/TheTuringPost/status/2022079178703847607

I looked into how Claude Code and Codex compare to the default scaffolds METR uses for time horizon measurements. It looks like they don’t significantly outperform our default scaffolds on any models we’ve tried them on so far.”” https://x.com/nikolaj2030/status/2022398669337825737

Dario acknowledges the multi-trillion dollar robotics opportunity, yet Anthropic is not hiring robotics talent; even as OpenAI and Google DeepMind aggressively build out their own robotics teams.”” https://x.com/TheHumanoidHub/status/2022416551270662427

EVMbench measures the ability of agents to detect, patch, and exploit smart contract vulnerabilities. Opus 4.6 getting mogged by GPT-5.2 and GPT-5.3. Although its detection accuracy is technically higher, it’s precision is much lower. (Opus is going shizo)”” https://x.com/scaling01/status/2024212205944643718

OpenAI’s Sam Altman and Anthropic’s Dario Amodei refuse to hold hands weeks after Super Bowl ad war | Fortune https://fortune.com/2026/02/19/openai-anthropic-sam-altman-dario-amodei-refused-to-hold-hands-ai-super-bowl-ad-war-ceos-big-tech-conflict/

OpenClaw creator on Opus vs Codex: “Opus is like the coworker that is a little silly sometimes, but it’s really funny and you keep him around. Codex is like the weirdo in the corner that you don’t want to talk to, but he’s reliable and gets shit done.” LMAO. Accurate.”” https://x.com/bilawalsidhu/status/2022571001490325791

Claude Sonnet 4.6 has landed #3 in Code and #13 in Text Arena! Highlights: ▪️+130 pts jump in Code Arena (#22 -> #3) compared to Sonnet 4.5, surpassing top-tier thinking models like Gemini-3.1 and GPT-5.2 ▪️Strong gains in Text categories: Math (#4) and Instruction Following”” https://x.com/arena/status/2024883614249615394

Use @Photoshop inside @ChatGPTapp now. No install required. It gives you slider-level control over effects like Cyber Glitch and Time Travel Warp. You can even open in Photoshop for web if you want to go deeper. #AdobePartner”” https://x.com/bilawalsidhu/status/2024178413511266802

“It is the first time I’ve seen AI solve a problem in my kind of theoretical physics that might not have been solvable by humans.” — Andy Strominger”” https://x.com/gdb/status/2022446135655436431

@sama I am very glad you got engaged deep into this experiment. Mathematical community needs strong signal from the AI labs that science is a serious engagement for you. Mathematics in its full proof-driven form is a pinnacle of human ingenuity and knowing how well the models can grasp”” https://x.com/nasqret/status/2022741621465403834

Advancing independent research on AI alignment | OpenAI https://openai.com/index/advancing-independent-research-ai-alignment/

Exclusive: OpenAI Has Poached Instagram’s Celebrity Whisperer | Vanity Fair https://www.vanityfair.com/news/story/openai-hires-charles-porch-instagram

feels like a significant milestone”” https://x.com/gdb/status/2022445080322449823

From OpenAI to Google, India hosts global AI summit | Reuters https://www.reuters.com/business/retail-consumer/openai-google-india-hosts-global-ai-summit-2026-02-16/

how did we ever write all that code by hand”” https://x.com/gdb/status/2022559457579012156

How will OpenAI compete? — Benedict Evans https://www.ben-evans.com/benedictevans/2026/2/19/how-will-openai-compete-nkg2x

I wonder what percentage of OAI employees understand the research question, the proof and the result of this paper”” https://x.com/scaling01/status/2022401147110318586

If you’re an infrastructure or security engineer, now is the best time to join OpenAI. It’s hard not to be inspired by what today’s coding tools are capable of, and we have line of sight to making them much better. While our core ML infrastructure problems remain much the same”” https://x.com/gdb/status/2023804170323849279

Instagram https://www.instagram.com/p/DU6K2tnkQKx/

it’s a good model”” https://x.com/gdb/status/2024611138760298999

OpenAI is now FedRamp 20x Low authorized. Excited to be on the path of bringing the most secure and powerful AI technologies to solve the hardest problems of some of our most important institutions. https://x.com/cryps1s/status/2024572447572582547

Prompt Caching 201 https://developers.openai.com/cookbook/examples/prompt_caching_201

Reverse-Engineering the OpenAI’s GPT-5 Tokenizer: What 200,000 Tokens Reveal About AEO/GEO | metehan.ai https://metehan.ai/blog/reverse-engineering-the-gpt-5-tokenizer-aeo-geo/

RLMs are exciting. GPT-5.2-Codex might be one of the better models for RLMs. These are tiny experiments, but exciting results so far. I have a few interesting ideas about how I want to use RLMs for code, long-context tasks, and large-scale analysis. More soon. Stay tuned!”” https://x.com/omarsar0/status/2024972027224846631

taste is a new core skill”” https://x.com/gdb/status/2023481258639286401

Thank you to our collaborators for their partnership. The preprint is available on arXiv and is being submitted for publication. We welcome feedback from the community.”” https://x.com/OpenAI/status/2022390104237707667

The weirdness in the early LLMs is not replicable using newer models, which is a shame sometimes. GPT-2 was objectively better at romantic candy hearts that got your attention.”” https://x.com/emollick/status/2023596249769808054

Using the long-standing Metacalculus bet of when a “”weakly general artificially intelligence is achieved””: ✅Loebner prize was a weak Turing Test, equivalent achieved by GPT-4.5 ✅Winograd passed by GPT-3 ✅SAT passed at 75% by GPT-4 All that’s left is a classic Atari game…”” https://x.com/emollick/status/2022519188212900036

Very excited about the “”First Proof”” challenge. I believe novel frontier research is perhaps the most important way to evaluate capabilities of the next generation of AI models. We have run our internal model with limited human supervision on the ten proposed problems. The”” https://x.com/merettm/status/2022517085193277874

Would really love to see that happening”” https://x.com/TheTuringPost/status/2022124372115566684

GPT-5.3-Codex-Spark is launching today as a research preview for Pro. More than 1000 tokens per second! There are limitations at launch; we will rapidly improve.”” https://x.com/sama/status/2022011797524582726

Introducing the Codex app | OpenAI https://openai.com/index/introducing-the-codex-app/

It’s probably one of the best places to start with LLMs in practice. The full core of GPT fits in 243 lines, then you just scale and optimize on top of it. This kind of share is a real gift.”” https://x.com/TheTuringPost/status/2023348280961495396

Wow gpt 5.3 Codex is actually so good. It has significantly better taste for UI design. My bet is that it’ll be #1 on @designarena once API is available.”” https://x.com/tkkong/status/2022410732760117403

A few thoughts after reading the @OpenAI paper on scattering amplitudes over the weekend: – The result from GPT-5.2 Pro is technically impressive, clearly at the level of a strong grad student / postdoc – However, without the transcripts, it is unclear what role the human”” https://x.com/_lewtun/status/2023334667064099207

GPT-5.2 derived a new result in theoretical physics. We’re releasing the result in a preprint with researchers from @the_IAS, @VanderbiltU, @Cambridge_Uni, and @Harvard. It shows that a gluon interaction many physicists expected would not occur can arise under specific”” https://x.com/OpenAI/status/2022390096625078389

Making progress in Quantum Field Theory with GPT-5.2. It’s happening, for real.”” https://x.com/SebastienBubeck/status/2022439681573695638

OpenAI CEO Sam Altman and Anthropic CEO Dario Amodei visibly declined to hold hands during a group photo at the India AI Impact Summit, even as other leaders on stage linked arms for the ceremonial shot”” https://x.com/Reuters/status/2024401067228684396?s=20

OpenAI and Broadcom announce strategic collaboration to deploy 10 gigawatts of OpenAI-designed AI accelerators | OpenAI https://openai.com/index/openai-and-broadcom-announce-strategic-collaboration/

(45) OpenAI Town Hall with Sam Altman – YouTube https://www.youtube.com/live/Wpxv-8nG8ec?t=3560s

Microsoft’s AI Chief Targets AI Self-Sufficiency and OpenAI Independence https://winbuzzer.com/2026/02/13/microsoft-mustafa-suleyman-ai-self-sufficiency-openai-mai-models-xcxwbn/

Our recent preprint on gluon amplitudes has sparked a lot of discussion, so I want to share the backstory — including how AI helped crack a problem that had stumped us for a year. I’ll also be giving a public lecture at Harvard this week. Details at the end.”” https://x.com/ALupsasca/status/2023402422320926762

Leave a Reply

Trending

Discover more from Ethan B. Holland

Subscribe now to keep reading and get access to the full archive.

Continue reading