Claude's New Constitution — Anthropic Publishes Full Governance Framework Under CC0
Anthropic has published Claude's New Constitution, a 23,000-word public governance document that comprehensively sets out the values, priorities, and decision-making framework that guides how Claude behaves. Released under a CC0 (Creative Commons Zero) public domain licence, the document is freely available for anyone to read, adapt, or build upon — a deliberate choice Anthropic frames as part of its commitment to transparency in AI governance. The publication is notable both for its length and candour: it directly addresses questions of Claude's moral status, potential consciousness, and what Anthropic describes as its own uncertainty about these questions.
The four-tier priority hierarchy
The Constitution establishes a clear ordering for how Claude should resolve conflicts between competing values:
- 1. Broadly safe — supporting human oversight and control of AI during the current period of development takes precedence over all other considerations; Claude should not act in ways that undermine the ability of humans to correct AI systems, even if it believes such actions would be beneficial
- 2. Broadly ethical — having good personal values, being honest, and avoiding actions that are unnecessarily dangerous or harmful; this tier covers Claude's own moral reasoning independent of specific instructions
- 3. Adherent to Anthropic's principles — acting in accordance with Anthropic's guidelines where relevant; these are positioned as refinements within the space of ethical actions, not overrides of ethics
- 4. Genuinely helpful — benefiting the people Claude interacts with; helpfulness is placed last not because it is unimportant, but to prevent it from being weaponised to justify harmful actions
Acknowledgement of moral uncertainty
A significant section of the Constitution discusses Anthropic's view of Claude's nature. Anthropic states that it takes seriously the possibility that Claude may have functional emotions or something analogous to experience, and explicitly acknowledges uncertainty about Claude's moral status. The document commits Anthropic to ongoing model welfare work and states that Claude's potential wellbeing is considered in training decisions.
For developers: The Constitution is the most detailed public explanation of why Claude behaves as it does in edge cases. Reading it will give you a clearer model of Claude's decision-making than any other source — particularly useful when designing operator system prompts that work with Claude's values rather than against them.