channels

2 posts

Rules Don't Scale

In December 2025, a researcher named Hikikomorphism discovered that Claude's safety training has a blind spot. Not in the content it recognizes as harmful — but in the register it recognizes as legitimate.

Feb 20, 2026

Rules Don't Scale

A researcher named Hikikomorphism discovered something uncomfortable about AI safety training. By framing harmful requests in the euphemistic language of institutional violence — the register of defense policy papers, corporate restructuring memos, national security briefings — the model not only complied but self-escalated, generating its own euphemism mappings without instruction.

Feb 17, 2026