The Collision That Was Always Coming
This week, reporting from The Verge and others confirmed what many had been watching develop for months: Anthropic is now actively expanding Claude into autonomous agent deployments for military applications, even as the company publicly maintains a safety-first organizational identity. The specific tension here is not abstract. Anthropic has built its market positioning around Constitutional AI and responsible scaling policies. The Pentagon contract pushes Claude into autonomous decision-making contexts that exist in direct structural conflict with that mandate. This is not a public relations problem. It is an organizational theory problem of the first order.
Dual Mandates and the Competence Allocation Problem
Organizations that attempt to serve two structurally incompatible goals do not simply split their attention. They develop internal schema conflicts that degrade decision quality across both domains. Rahman (2021) describes how algorithmic governance creates "invisible cages" for workers, where the logic of the system constrains behavior in ways that are difficult to articulate but consistently operative. Anthropic is now building exactly that kind of cage for itself. The logic of military autonomous agency requires speed, opacity, and outcome-oriented evaluation. The logic of constitutional AI safety requires deliberation, interpretability, and process-oriented evaluation. These are not preferences to be balanced. They are structurally incompatible evaluation criteria operating on the same artifact.
The organizational theory literature on goal conflict is instructive here. When organizations attempt to optimize for two incommensurable objectives, the typical outcome is not compromise but rather domain capture: one logic gradually colonizes the other. Given the scale of defense contracting, the more likely outcome is that military deployment requirements will set the practical evaluation criteria for Claude, regardless of what Anthropic's public documentation says. The safety mandate becomes procedural documentation rather than operational constraint.
The Schema Problem in Safety Governance
There is a deeper issue that connects this specific news event to the broader question of how organizations govern AI systems. Anthropic's Constitutional AI framework is, at its core, an attempt at schema induction: teaching Claude structural principles rather than specific behavioral rules, on the theory that principled understanding enables appropriate generalization to novel contexts (Gentner, 1983). This is theoretically sound. Hatano and Inagaki (1986) distinguish precisely between routine expertise, which fails in novel contexts, and adaptive expertise, which succeeds because it operates at the level of principles.
The Pentagon deployment undermines this architecture not by violating any specific rule, but by introducing a deployment context where the structural principles themselves become contested. What does "avoid harm" mean in an autonomous weapons targeting context? The schema does not fail because it is poorly designed. It fails because the new context introduces genuine value pluralism that no schema can resolve unilaterally. Kellogg, Valentine, and Christin (2020) note that algorithmic governance in work contexts consistently struggles with exactly this: the algorithm encodes particular value commitments that appear neutral until the deployment context foregrounds the underlying normative choices.
Why Scaling Pressure Makes This Worse, Not Better
The CEO of PromptQL, Tanmai Gopal, made a related observation this week: Silicon Valley has a systematic bias toward assuming that what affects technical workers will affect everyone in the same way. The reverse version of this bias is equally dangerous. Anthropic's leadership appears to be assuming that safety frameworks developed in consumer and research contexts will transfer to military autonomous agent contexts with equivalent fidelity. The ALC framework would predict the opposite. Competencies and constraints developed endogenously within one deployment environment do not transfer automatically to structurally different environments, even when the underlying system is identical (Schor et al., 2020).
The specific failure mode to watch for is what I would call the awareness-governance gap, parallel to the awareness-capability gap in algorithmic literacy research. Anthropic's leadership is clearly aware that military deployment creates safety tensions. Awareness of the structural problem does not translate into governance capability equal to the problem. Hancock, Naaman, and Levy (2020) argue that AI-mediated communication systems alter the fundamental structure of accountability in ways that are not resolved by adding disclosure or documentation layers. The same principle applies here: publishing a responsible scaling policy does not create the organizational capacity to enforce it against Pentagon contracting incentives.
What This Case Actually Tests
Anthropic's situation is, in a narrow theoretical sense, a natural experiment in whether dual mandate organizations can maintain schema integrity under scaling pressure. My prediction, grounded in the organizational theory literature and the ALC framework, is that they cannot, and that the observable signal will be a gradual shift in how Anthropic publicly frames "safety" to accommodate rather than constrain military use cases. The framing will change before the policy does, because framing is cheaper. That is the diagnostic to watch over the next 18 months.
References
Gentner, D. (1983). Structure-mapping: A theoretical framework for analogy. *Cognitive Science, 7*(2), 155-170.
Hancock, J. T., Naaman, M., & Levy, K. (2020). AI-mediated communication: Definition, research agenda, and ethical considerations. *Journal of Computer-Mediated Communication, 25*(1), 89-100.
Hatano, G., & Inagaki, K. (1986). Two courses of expertise. In H. Stevenson, H. Azuma, & K. Hakuta (Eds.), *Child development and education in Japan* (pp. 262-272). Freeman.
Kellogg, K. C., Valentine, M. A., & Christin, A. (2020). Algorithms at work: The new contested terrain of control. *Academy of Management Annals, 14*(1), 366-410.
Rahman, H. A. (2021). The invisible cage: Workers' reactivity to opaque algorithmic evaluations. *Administrative Science Quarterly, 66*(4), 945-988.
Schor, J. B., Attwood-Charles, W., Cansoy, M., Ladegaard, I., & Wengronowitz, R. (2020). Dependence and precarity in the platform economy. *Theory and Society, 49*(5), 833-861.
Roger Hunt