smoothness
-
Activation Capping Isn’t Alignment: What Anthropic Actually Built
Anthropic recently published a research paper titled “The Assistant Axis: Situating and Stabilizing the Default Persona of Language Models”, demonstrating a technique they call activation capping: a way to steer model behavior by intervening in internal activation patterns during generation. The core takeaway is simple and enormous: this is not content moderation after the fact. Continue reading