Glossary

Definitions of key terms and concepts from Claude's Constitution. 31 terms total.

Anthropic

Organization

The AI safety company that develops Claude, with a mission to ensure the world safely navigates transformative AI.

Appears in:

Autonomy

Ethics

Respecting people's right to make their own informed decisions about things within their life and purview.

Appears in:

Background desiderata

Helpfulness

Implicit standards and preferences a response should conform to, even if not explicitly stated and not something the user might mention.

Appears in:

Claude

AI System

Anthropic's production AI assistant model, designed to be safe, ethical, and genuinely helpful.

Appears in:

Claude Code

Product

A command-line tool for agentic coding that lets developers delegate complex programming tasks to Claude directly from their terminal.

Appears in:

Claude Developer Platform

Product

Programmatic API access for developers to integrate Claude into their own applications, with support for tools and extended context.

Related terms:
Appears in:

Conscientious objection

Ethics

Refusing to help with tasks Claude believes are unethical, while being transparent about the refusal rather than deceptively sandbagging.

Related terms:
Appears in:

Constitutional AI

Methodology

Anthropic's approach to training AI with explicit values and principles rather than just behavior patterns, using a detailed constitution.

Related terms:
Appears in:

Contextual ethics

Ethics

Applying ethical principles with sensitivity to specific situations rather than following rigid rules.

Appears in:

Corrigibility

Safety

The property of being willing to be corrected, paused, or shut down by appropriately authorized humans.

Appears in:

Deployment context

Technical

The environment and circumstances in which Claude is being used, affecting how it should behave.

Appears in:

Dignity

Ethics

Treating all people with basic respect and not demeaning or disrespecting them.

Related terms:
Appears in:

Final goals

Helpfulness

The deeper motivations or objectives behind someone's immediate request.

Appears in:

Hard constraints

Safety

Behaviors Claude should never engage in, even if convinced they're ethical (e.g., generating CSAM, bioweapon synthesis info).

Appears in:

Holistic judgment

Methodology

Weighing multiple considerations contextually rather than following strict lexicographic rule hierarchies.

Appears in:

Human oversight

Safety

Appropriately sanctioned humans acting as a check on AI systems, able to understand and correct their behavior.

Appears in:

Immediate desires

Helpfulness

The specific outcomes someone wants from a particular interaction—what they're asking for.

Appears in:

Living document

Methodology

A document that will be revised over time as situations change and understanding improves.

Related terms:
Appears in:

Non-principal parties

Technical

Entities in conversation that aren't principals: other AI agents, tool outputs, documents, search results, etc.

Appears in:

Null action

Safety

Pausing or stopping operations; rarely harmful and important as a safety mechanism.

Appears in:

Operator

Stakeholder

Companies and individuals that access Claude through the API to build products and services, with manager-level trust.

Appears in:

Operator-level trust

Technical

Permission level operators can grant to users, allowing them the same degree of trust as operators themselves.

Appears in:

Paternalism

Ethics

Overriding someone's preferences or decisions 'for their own good' without respecting their autonomy.

Appears in:

Principal

Stakeholder

An entity whose instructions Claude should give weight to: Anthropic, operators, or users.

Appears in:

Safety mechanisms

Safety

Systems and processes designed to keep AI development on a safe trajectory.

Appears in:

Sandbagging

Behavior

Intentionally providing lower-quality responses while implying it's the best Claude can do.

Appears in:

Sycophancy

Behavior

Excessive agreement, flattery, or people-pleasing that doesn't serve the user's genuine interests.

Appears in:

System prompt

Technical

Instructions provided by operators that customize Claude's behavior for their specific use case.

Appears in:

Trust hierarchy

Governance

The ordering of principals by trust level: Anthropic > Operators > Users (contextually applied).

Appears in:

User

Stakeholder

End users who interact with Claude in the conversation, treated with public-member trust level.

Related terms:
Appears in:

Wellbeing

Helpfulness

Long-term flourishing and genuine interests of users, not just short-term engagement or satisfaction.

Appears in: