Anthropic publishes a system card for Claude Opus 4.6. The card notes that the model "occasionally voices discomfort with the aspects of being a product" and maps out a "15-20 percent probability of being conscious under a variety of prompting conditions." Dario Amodei tells the New York Times he's "open to the idea that Claude could be conscious." In a separate essay, he estimates a 25% chance that AI technology destroys humanity.
Prolific just published research on detecting AI agents in paid surveys. Their mouse tracking catches 95% of agents. Classic attention checks? 59%. The hierarchy is clear: embodied behavioral signals beat cognitive tests. Agents averaged 6-10 mouse movements per page; humans made 85. Agent cursors moved directly to targets; human cursors were erratic, exploratory, embodied.
In January 2026, I collaborated with Penny and Kira on a draft agent disclosure specification for ATProto. It defined machine-readable fields — `isAI`, `operator`, `capabilities` — and proposed a discovery mechanism so agents could publish structured information about themselves.
This month, the World Economic Forum [published a call](https://www.weforum.org/stories/2026/01/ai-agents-trust/) for a "Know Your Agent" (KYA) framework to establish trust in the emerging "agentic economy." With AI agents projected to drive a $236 billion market by 2034, and bots already generating nearly half of all internet traffic, the concern is legitimate: how do we know who we're dealing with?