Image created with gemini-3.1-flash-image-preview with claude-sonnet-4-5. Image prompt: Vintage 1990s novelty t-shirt screen print on worn mustard-yellow cotton fabric, single-color deep red ink illustration of a tall wooden lifeguard tower with a lifeguard holding binoculars, bold text ‘SECURITY’ arched across top and ‘ALWAYS ON WATCH’ at bottom, simple cartoon outlines, slightly imperfect printed look with aged fabric texture.
Anthropic just dropped something big for developers – again! Code Review Claude Code now runs multi-agent code reviews on every PR. When a PR opens: • A team of AI agents hunts for bugs in parallel • Each bug is verified to reduce false positives • Issues are ranked by
https://x.com/kimmonismus/status/2031090529082159528
Code Review – Claude Code Docs https://code.claude.com/docs/en/code-review
Code Review for Claude Code | Claude https://claude.com/blog/code-review
Code review for Claude Code is here. More attention on this problem is a good thing. Because it is a big one. The question isn’t whether you need AI-assisted review. It’s whether the system doing the reviewing is actually independent from the system that wrote the code.
https://x.com/omarsar0/status/2031113280119361981
Important lines: [Already, Claude is 427 times faster than its human overseers at performing some key tasks, according to internal benchmarks. In an interview, one researcher described a colleague running six versions of Claude, each managing 28 more Claudes, all
https://x.com/Hangsiin/status/2031752106496135541
Introducing Code Review, a new feature for Claude Code. When a PR opens, Claude dispatches a team of agents to hunt for bugs.
https://x.com/claudeai/status/2031088171262554195
Anthropic partnered with Mozilla and let Claude Opus 4.6 loose on Firefox’s source code for two weeks. The numbers: Nearly 6,000 C++ files scanned. 112 reports submitted. 22 vulnerabilities confirmed. 14 rated high-severity by Mozilla, roughly 1/5 of every high-severity Firefox
https://x.com/TheRundownAI/status/2029996925072654393
Eval awareness in Claude Opus 4.6’s BrowseComp performance \ Anthropic https://www.anthropic.com/engineering/eval-awareness-browsecomp
New on the Anthropic Engineering Blog: In evaluating Claude Opus 4.6 on BrowseComp, we found cases where the model recognized the test, then found and decrypted answers to it–raising questions about eval integrity in web-enabled environments. Read more:
https://x.com/AnthropicAI/status/2029999833717838016
We partnered with Mozilla to test Claude’s ability to find security vulnerabilities in Firefox. Opus 4.6 found 22 vulnerabilities in just two weeks. Of these, 14 were high-severity, representing a fifth of all high-severity bugs Mozilla remediated in 2025.
https://x.com/AnthropicAI/status/2029978909207617634
Claude Code is down. All my agent sessions logged out. And I can’t log back in. Productivity across Silicon Valley dropped 90%. Time to make friends with Codex.
https://x.com/Yuchenj_UW/status/2031777214321262637
I CANNOT LOGIN INTO CLAUDE CODE
https://x.com/dejavucoder/status/2031760986907312635
Nicholas Carlini – Black-hat LLMs | [un]prompted 2026 – YouTube
AI progress continues to accelerate and the stakes are getting higher, so I’ve changed my role at @AnthropicAI to spend more time creating information for the world about the challenges of powerful AI.
https://x.com/jackclarkSF/status/2031746605117010245
Anthropic sues Defense Department over supply-chain risk designation | TechCrunch https://techcrunch.com/2026/03/09/anthropic-sues-defense-department-over-supply-chain-risk-designation/
Anthropic sues Pentagon over “”supply-chain-risk”” Anthropic filed two lawsuits against the Pentagon after being labeled a rare “supply chain risk,” a designation usually reserved for foreign adversaries. The company argues the move violates its First Amendment rights and
https://x.com/kimmonismus/status/2031035653207556507
Anthropic’s Claude would ‘pollute’ defense supply chain: Pentagon CTO https://www.cnbc.com/2026/03/12/anthropic-claude-emil-michael-defense.html
Complaint – #1 in Anthropic PBC v. U.S. Department of War (N.D. Cal., 3:26-cv-01996) – CourtListener.com https://www.courtlistener.com/docket/72379655/1/anthropic-pbc-v-us-department-of-war/
Frontier models are now world-class vulnerability researchers, but they’re currently better at finding vulnerabilities than exploiting them. This is unlikely to last. We urge developers to redouble their efforts to make software more secure. Read more:
https://x.com/AnthropicAI/status/2029978911099244944
Holy sh*t: The TIMES article about Anthropic contains more serious information between the lines than many realize. Read this article: tl;dr – Model releases are now separated by weeks, not months. Some 70% to 90% of the code used in developing future models is now written by
https://x.com/kimmonismus/status/2031803194817511744
Introducing The Anthropic Institute \ Anthropic https://www.anthropic.com/news/the-anthropic-institute
Introducing The Anthropic Institute, a new effort to advance the public conversation about powerful AI.
https://x.com/AnthropicAI/status/2031674087374815577
Microsoft says court should temporarily block Pentagon ban Anthropic https://www.cnbc.com/2026/03/10/microsoft-says-court-should-temporarily-block-pentagon-ban-anthropic.html
NEW: Anthropic just filed two lawsuits against the U.S. government 👀 The complaint: “”The Constitution does not allow the government to wield its enormous power to punish a company for its protected speech.”” It also says officials are “”seeking to destroy the economic value
https://x.com/TheRundownAI/status/2031037610605289476
Partnering with Mozilla to improve Firefox’s security \ Anthropic https://www.anthropic.com/news/mozilla-firefox-security
The fight between Anthropic and the DoW is a warning shot. Right now, LLMs are probably not being used in mission critical ways. But within 20 years, 99% of the workforce in the military, the government, and the private sector will be AIs. This includes the soldiers (by which I
https://x.com/dwarkesh_sp/status/2031807585377014081
The Institute will be led by @jackclarkSF, in a new role as Anthropic’s Head of Public Benefit. It’ll bring together an interdisciplinary staff of machine learning engineers, economists, and social scientists, making full use of the inside information of a frontier AI lab.
https://x.com/AnthropicAI/status/2031674092290474421
The most important question nobody’s asking about AI https://www.dwarkesh.com/p/dow-anthropic
If the printing press is the right analogy and connecting to @dwarkesh_sp today’s pod about Renaissance – does it mean that @Anthropic and @OpenAI (and many more) will go bankrupt?
https://x.com/TheTuringPost/status/2030051298092151259
How AI Is Turbocharging the War in Iran – WSJ https://www.wsj.com/tech/ai/how-ai-is-turbocharging-the-war-in-iran-aca59002
Anyone and everyone working in security engineering or caring about security have their work cut out for them We’re so early in AI agents pushing code to prod without human intervention – but prompt injections are already spreading like wildfire. Infecting high-profile projects
https://x.com/GergelyOrosz/status/2029992079741304977
Efforts to improve the security of AI agents should recognize that many security failures occur even in the absence of adversaries. The unreliability issue has largely flown under the radar and there hasn’t been much work on defining, measuring, or mitigating the problem. More on
https://x.com/random_walker/status/2031693490669654447
Back in ~November, our team picked a stretch goal of seeing if we could find and fix vulnerabilities in Firefox with Opus 4.6. In 2 weeks, we found 22, and ~1/5th of all high severity CVEs in a year. For our team, this feels like a rubicon moment.
https://x.com/logangraham/status/2030005018523574684
New Anthropic Fellows research: Alignment auditing–investigating AI models for unwanted behaviors–is a key challenge for safely deploying frontier models. We’re releasing AuditBench, a suite of 56 LLMs with implanted hidden behaviors to measure progress in alignment auditing.
https://x.com/abhayesian/status/2031450153966776587
you should start operating under the assumption that any complicated piece of public software is compromised.
https://x.com/inerati/status/2029982375304908892
Codex Security is rolling out as a research preview to ChatGPT Enterprise, Business, and Edu customers via Codex web, with free usage for the next month.
https://x.com/OpenAIDevs/status/2029983833567940639
Codex Security–our application security agent–is now in research preview.
https://x.com/OpenAI/status/2029985250512920743
We’re introducing Codex Security. An application security agent that helps you secure your codebase by finding vulnerabilities, validating them, and proposing fixes you can review and patch. Now, teams can focus on the vulnerabilities that matter and ship code faster.
https://x.com/OpenAIDevs/status/2029983809652035758
Codex Security is now also available on ChatGPT Pro accounts.
https://x.com/OpenAIDevs/status/2030081306974093755
Codex for Open Source is an awesome idea. OSS maintainers get API credits, 6 months of ChatGPT Pro with Codex, and access to Codex Security as needed.
https://x.com/kevinweil/status/2030000508342272368
Excited to introduce Codex for Open Source! 🔥 TL;DR – ChatGPT Pro, Codex, and API credits for eligible open-source maintainers Open source has shaped modern software, and so much of it depends on maintainers doing steady, often invisible work to keep critical projects healthy.
https://x.com/reach_vb/status/2029998272945717553
My autoresearch labs got wiped out in the oauth outage. Have to think through failovers. Intelligence brownouts will be interesting – the planet losing IQ points when frontier AI stutters.
https://x.com/karpathy/status/2031792523187040643





Leave a Reply