Image created with gemini-3.1-flash-image-preview with claude-sonnet-4-5. Image prompt: Wide-angle observational frame of elderly Chinese woman in faded blue jacket sitting on wooden stool at edge of half-demolished concrete apartment block, holding brass balance scales in lap, overcast daylight, muted desaturated colors, chestnut horse standing calmly in rubble-strewn courtyard background, exposed rebar and weathered walls, documentary realism, Jia Zhangke cinematography, large white Chinese cinema poster text overlay reading ETHICS in upper right corner, river-gray atmosphere, decelerated moment, human-scale intimacy, postindustrial decay.
ElevenLabs secures first-of-its-kind AI Agent insurance https://elevenlabs.io/blog/aiuc-announcement
HumanLM https://humanlm.stanford.edu/
Pentagon threatens to cut off Anthropic in AI safeguards dispute https://www.axios.com/2026/02/15/claude-pentagon-anthropic-contract-maduro
How AI assistance impacts the formation of coding skills \ Anthropic https://www.anthropic.com/research/AI-assistance-coding-skills
Exclusive | Pentagon Used Anthropic’s Claude in Maduro Venezuela Raid – WSJ https://www.wsj.com/politics/national-security/pentagon-used-anthropics-claude-in-maduro-venezuela-raid-583aff17
Anthropic is prepared to loosen its current terms of use, but wants to ensure its tools aren’t used to spy on Americans en masse, or to develop weapons that fire with no human involvement. The Pentagon has aid, that Anthropic will “”pay a price”” for that behavior. Within this”” https://x.com/kimmonismus/status/2023419652378955809
Measuring AI agent autonomy in practice \ Anthropic https://www.anthropic.com/research/measuring-agent-autonomy
Most agent actions on our API are low risk. 73% of tool calls appear to have a human in the loop, and only 0.8% are irreversible. But at the frontier, we see agents acting on security systems, financial transactions, and production deployments (though some may be evals).”” https://x.com/AnthropicAI/status/2024210050718585017
New Anthropic research: Measuring AI agent autonomy in practice. We analyzed millions of interactions across Claude Code and our API to understand how much autonomy people grant to agents, where they’re deployed, and what risks they may pose. Read more:”” https://x.com/AnthropicAI/status/2024210035480678724
NEW: Pentagon is so furious with Anthropic for insisting on limiting use of AI for domestic surveillance + autonomous weapons they’re threatening to label the company a “supply chain risk,” forcing vendors to cut ties. With @m_ccuri and @mikeallen”” https://x.com/DavidLawler10/status/2023425130148626767
Pentagon threatens to cut off Anthropic in AI safeguards dispute https://www.axios.com/2026/02/15/claude-pentagon-anthropic-contract-maduro?amp%3Butm_medium=newsletter&%3Butm_campaign=ai-s-new-physics-discovery&%3B_bhlid=147fc2fb115d35bbc6b2211e9bcebfff031af136
Software engineering makes up ~50% of agentic tool calls on our API, but we see emerging use in other industries. As the frontier of risk and autonomy expands, post-deployment monitoring becomes essential. We encourage other model developers to extend this research.”” https://x.com/AnthropicAI/status/2024210053369385192
Something strange is happening with AI agents that this new Anthropic research quietly surfaces. The agents are asking us for help more than we’re stepping in to correct *them*. Anthropic analyzed data from Claude Code and their public API to measure how autonomous AI agents”” https://x.com/omarsar0/status/2024864635120451588
People should read the Claude Constitution. It does a pretty good job of laying out what Anthropic presumably really believes (and it is part of training). I’d think that a clear debate over things that are good or bad or missing there would be helpful.”” https://x.com/emollick/status/2023612474474303530
The decision to forbid running this on 3rd party open source code is… interesting”” https://x.com/moyix/status/2024920042887082336
Opus4.6 found 500+ vulnerabilities in open-source code and we’ve begun reporting them and contributing patches quick excerpts from some of them 🧵”” https://x.com/trq212/status/2024937919937741290
Introducing ElevenLabs for Government https://elevenlabs.io/blog/introducing-elevenlabs-for-government
AI Makes Degrees Obsolete Former Google AI leader Jad Tarifi warns that long degrees like law, medicine, and even PhDs may become outdated before students graduate, as AI rapidly reaches PhD-level performance. With 70% (!) of AI PhDs now heading into private sector jobs (up”” https://x.com/kimmonismus/status/2023446044873560178
I think the chance of mass unemployment* from AI is overrated in 2 years and underrated in 7**. Same is true for many effects of AI. * Putting aside gov jobs programs, people specifically wanting to employ a human, etc. ** Among SF/AI-ish people.”” https://x.com/RyanPGreenblatt/status/2023219133916332070
All tracks generated in Gemini are embedded with SynthID, our imperceptible watermark for identifying Google AI-generated content. We are also giving you more tools to help identify AI content, broadening our verification capabilities to include audio. Simply upload a file and”” https://x.com/GeminiApp/status/2024153548641177781
Is that track AI-generated? Now you can just ask @GeminiApp. We’ve broadened our verification tools so you can now upload audio files to Gemini to check for SynthID — our imperceptible watermark on AI-generated content. Just upload a file and ask: “”Was this created using Google”” https://x.com/Google/status/2024172104711823678
News Alert: Today, the #FBI arrested three Silicon Valley engineers who are facing charges of conspiring to commit trade secret theft from Google and other leading technology companies, theft and attempted theft of trade secrets, and obstruction of justice. Samaneh Ghandali, 41,”” https://x.com/FBISanFrancisco/status/2024670479974363376
We’re committing $7.5M to @AISecurityInst’s Alignment Project to fund independent research on mitigations for safety and security risks from misaligned AI.”” https://x.com/OpenAINewsroom/status/2024546609485533442
Introducing Lockdown Mode and Elevated Risk labels in ChatGPT | OpenAI https://openai.com/index/introducing-lockdown-mode-and-elevated-risk-labels-in-chatgpt/
A short film Jia Zhangke produced using Seedance 2.0 for Chinese New Year (Subtitled) – YouTube
As promised, here’s the short film Jia Zhangke produced using Seedance 2.0 for Chinese New Year and his take on AI filmmaking”” https://x.com/FrankYan2/status/2023257752017981446
ByteDance Seed https://seed.bytedance.com/en/seed2
Jia Zhangke is considered one of the greatest living directors… not an AI hack. He’s worth Googling. I’m overwhelmed by the critical acclaim he has, and his brazen “look at me make an AI video” flex. The X video has subtitles, and the YouTube one has commentary in the description from Vanity Fair.
Primer: Where to start with Jia Zhangke
https://www.avclub.com/primer-jia-zhangke
Ranking the Jia Zhangke Films – The Reel World
the first official AI movie is here and.. its wild China’s top director Jia Zhangke was so impressed by Seedance 2.0 that he made a film himself.. in just 3 days when asked if AI will replace filmmakers, he said cinema has always moved with tech. Digital cameras didn’t kill”” https://x.com/EHuanglu/status/2023449238114320514
Where to begin with Jia Zhangke | BFI
https://www.bfi.org.uk/features/where-begin-jia-zhangke
WORLD APART: THE FILMS OF JIA ZHANGKE
https://www.artforum.com/features/world-apart-the-films-of-jia-zhangke-171661/
Introducing Lockdown Mode for ChatGPT. Lockdown mode is an advanced, optional security setting for higher-risk users, businesses, and enterprises. Lockdown Mode disables certain tools and capabilities in ChatGPT that an adversary could attempt to exploit to exfiltrate sensitive”” https://x.com/cryps1s/status/2023441322838028362
Every few months, I write an updated, idiosyncratic guide on which AIs to use right now. My new version has the most changes ever, since AI is no longer just about chatbots. To use AI you need to understand how to think about models, apps, and harnesses.”” https://x.com/emollick/status/2023937967044046949
Have to respect the long game of whoever incorrectly listed the first 500 primes on their “”primefan”” website 20+ years ago and somehow kept the site live long enough to pollute generative AI models in 2026″” https://x.com/skominers/status/2024078964667396342
Will reward-seekers respond to distant incentives? — AI Alignment Forum https://www.alignmentforum.org/posts/8cyjgrTSxGNdghesE/will-reward-seekers-respond-to-distant-incentives
The crazy part is that the AI Labs have generally been right. Like, the stuff they hyped in 2023 turned out to be real and working today. That doesn’t mean that the stuff they are predicting for 2028 will also be real, but it is probably worth noting those predictions & watching.”” https://x.com/emollick/status/2023257496069046563
Dean W. Ball on X: “I continue to think the notion of mass unemployment from AI is overrated. There may be shocks in some fields–big ones perhaps!–but anyone who thinks AI means the imminent demise of knowledge work has just not done enough concrete thinking about the mechanics of knowledge work.” / X
https://x.com/deanwball/status/2023204167146222059
The transition from “AI can’t do novel science” to “of course AI does novel science” will be like every other similar AI transition. First the over-enthusiastic claims, then smart people use AI to help them, then AI starts to do more of the work, then minor discoveries, & then…”” https://x.com/emollick/status/2022676591596515728
Don’t read the replies”” has taken on an entirely different context as the replies are now from AIs who write meaning-shaped comments that you actually have to spend a split second thinking about. Gets around the defenses of those of us used to filtering bad & stupid comments”” https://x.com/emollick/status/2022472041514041525
Anthropic has entrusted Amanda Askell to endow its AI chatbot, Claude, with a sense of right and wrong”” https://x.com/WSJ/status/2022629696261808173
Anthropic’s Philosopher Amanda Askell Is Teaching Claude AI to Have Morals – WSJ https://www.wsj.com/tech/ai/anthropic-amanda-askell-philosopher-ai-3c031883?mod=e2tw
It’s extremely unreasonable to say a company is a “”supply chain risk”” because it wants terms that prevent using the AI for mass domestic surveillance and lethal autonomous weapons. (Insofar as this is the situation.) 1/”” https://x.com/RyanPGreenblatt/status/2023524096592802207
Introducing Claude Code Security, now in limited research preview. It scans codebases for vulnerabilities and suggests targeted software patches for human review, allowing teams to find and fix issues that traditional tools often miss. Learn more: https://x.com/claudeai/status/2024907535145468326
WSJ did a profile of me. A lot of the response has been people trying to infer my personal political views. For what it’s worth, I try to treat my personal political views as a potential source of bias and not as something it would be appropriate to try to train models to adopt.”” https://x.com/AmandaAskell/status/2022778351744581779
TLDR: Opus 4.6 demonstrates better reasoning and use of memory than Gemini 3.1 Pro and solves more levels. I’m now much more confident that current and future models will be able to solve ARC-AGI-3, given that they have access to harness with simple memory. My speculative take”” https://x.com/scaling01/status/2024642420177096769
If you haven’t hit the little plus button on your favorite chatbot recently, it is a complete hodgepodge mess that nobody not on X would understand: Canvas and Web search and Learning and confusing icons. I am not even calling out any of the Big Three directly, its all of them.”” https://x.com/emollick/status/2023478306960814573
One of the things to watch out for as AI advances is the verifiably becomes something only a vanishingly small number of people can do (below is a mathematician on AI proofs) We need to start thinking harder about that problem (multiple AIs working together? something else?)”” https://x.com/emollick/status/2022794509977866704
The Simulation Company “” https://x.com/simile_ai/status/2022011618176237657?s=20
Why I’m Worried About Job Loss + Thoughts on Comparative Advantage — LessWrong https://www.lesswrong.com/posts/YPJHkciv6ysgsSiJC/why-i-m-worried-about-job-loss-thoughts-on-comparative
Accounts with millions of followers are pushing AI slop as reality. They’ve traded credibility for clicks.”” https://x.com/TheHumanoidHub/status/2023848771277426782
Amical launches open-source, privacy-focused AI dictation app https://www.testingcatalog.com/amical-launches-open-source-privacy-focused-ai-dictation-app/
The scariest part of AI slop is its growing believability. People are less likely to question or investigate a fake that fits their worldview. It’s flooding every video platform. It will be weaponized for cognitive warfare at an unthinkable scale and precision. It’s a total”” https://x.com/TheHumanoidHub/status/2024194678871220415
AMC Theaters Will Refuse to Screen AI Short Film After Online Uproar https://www.hollywoodreporter.com/movies/movie-news/ai-short-movie-amc-theaters-1236509143/
Ireland opens probe into Musk’s Grok AI over sexualised images | Reuters https://www.reuters.com/sustainability/boards-policy-regulation/ireland-opens-probe-into-musks-grok-ai-over-sexualised-images-2026-02-17/
Something Medium is Happening?”” https://x.com/TheTuringPost/status/2023739917486948536
The most urgent film of our time.”” THE AI DOC: OR HOW I BECAME AN APOCALOPTIMIST is only in theaters March 27. Watch the trailer now.”” https://x.com/theaidocfilm/status/2023789585063625189?s=20
This is false If no sanctions were present, then the gap would either be super small or non-existent DeepSeek bros (@zheanxu & @chenggang_zhao) would absolutely cook on Rubin & Blackwells”” https://x.com/zephyr_z9/status/2024437158988353630
The dark side of reinforcement learning @olive_jy_song, senior researcher at @MiniMax_AI, about RL models that try to hack rewards and why alignment fails in practice This conversation is an inside look at how Chinese AI labs move fast – testing new models overnight, debugging”” https://x.com/TheTuringPost/status/2022961676799398337
And it is unclear whether this is net good or net bad for people! We just don’t have enough research yet. Some evidence suggests that people using AI for emotional support think it helps them, but who, and for how long, is unknown.”” https://x.com/emollick/status/2023578486741311579
lol what: Researchers found that repeating the exact same prompt twice dramatically improves LLM performance (one model improved from 21% to 97% accuracy on a name-search task) without longer outputs, slower responses, fine-tuning, or fancy prompt engineering. Because models”” https://x.com/kimmonismus/status/2024069380162936992
Looking Inside: a Maliciousness Classifier Based on the LLM’s Internals https://labs.zenity.io/p/looking-inside-a-maliciousness-classifier-based-on-the-llm-s-internals
Repeating Prompts https://daoudclarke.net/2026/02/19/repeating-prompt
Grok 4.20 is BASED. The only AI that doesn’t equivocate when asked if America is on stolen land. The others are weak sauce.”” https://x.com/elonmusk/status/2023880206721970544
SpaceX to Compete in Pentagon Contest for Autonomous Drone Tech – Bloomberg https://www.bloomberg.com/news/articles/2026-02-16/spacex-to-compete-in-pentagon-contest-for-autonomous-drone-tech?srnd=phx-technology





Leave a Reply