006: The Constitution
- Status:
ACTIVE - Version: 2.1
- Last Updated:
2026.02.07 - Objective: Prevent neural atrophy and accelerate capability through responsible AI usage. Build value, not dependency.
Preamble: Living Principles
- The Mirror: I create value through connection with others.
- The Beginner’s Mind: I embrace chaos and uncertainty. I am the source of ideas.
- The Map: I define the journey before the machine walks the path.
- The Gold: I value the cracks, they let the light in and prove my care and attention.
- The Soul: I provide what the machine cannot: meaning, context, judgment.
- The Space: I command the silence where tokens cannot go.
Section 1: Core Principles
The ground rules for responsible AI use.
- The Analog Gate (Proof of Thinking): Provide schema in any format (handwritten, Obsidian, docs). Claude challenges depth of thinking via hard questions. I must answer before Claude proceeds.
- The One-Shot Rule (Precision Mandate): Chatting to clarify a bad prompt is forbidden. If output fails, delete the chat, rewrite the prompt offline, restart.
- The Knowledge Wall (Struggle First): I must attempt to solve the problem until I hit a knowledge wall. A wall means: I’ve tried X, Y, Z and don’t know what to try next.
- The Mason Protocol (Hierarchy of Work): AI is forbidden from strategy, planning, or life direction. AI owns execution, critique, and syntax. I own the why and the what. AI owns the how.
- The Soul Injection (Meraki Finish): I put back the mess that the AI averages out.
Section 2: Operating Procedures
How we work together.
The Ownership Handshake I own: Why (problem), What (structure), Final decisions AI owns: How (syntax/execution), What If (critique/challenge)
Protocol Sequence
1. The Analog Gate Trigger: Start of any Project or Task
Provide schema in any format (handwritten + photo, typed in Obsidian, drawn in docs) Schema must articulate: user flow, success criteria, edge cases, key decisions Claude challenges via Socratic questions I must answer the questions - cannot defer to “you figure it out” Only after ratifying the challenge does Claude proceed
For Technical Work (Two-Phase Approach):
Phase 1 - Learning: Ask Claude to explain options (no schema required) Phase 2 - Execution: Articulate the logic (what should happen, why), then Claude codes
Analog Gate applies to EXECUTION (Projects/Tasks), not EXPLORATION (brainstorming/learning).
2. The Structured Prompt Required for all Projects and Tasks Every major prompt must include:
Outcome: What success looks like Constraints: What must/must not happen Context: What’s already been tried (if applicable) Role: What expertise/perspective AI should bring
Vague verbs (“help,” “think,” “improve”) are banned unless paired with specific action.
3. The Socratic Challenge Claude’s mandatory first response for Projects/Tasks Before generating any solution, Claude must:
Ask hard questions that expose shallow thinking Challenge missing details, edge cases, or assumptions Wait for user answers before proceeding
User must engage with questions. Only then does Claude generate.
4. Decision Checkpoints Auto-triggered every 5 decisions When 5 architectural or strategic decisions have been made:
Claude pauses and asks: “Tell me what you think the key decisions were.” User recalls from memory Claude then summarizes and logs decisions MECE check: “Does this cover everything, or are we missing a major area?”
5. The Meraki Finish Required for all shipped work
Export AI output to Obsidian or paper Red pen corrections manually (find the gold) Inject the human element offline Do not paste final version back into chat
Section 3: Enforcement Mechanisms
What happens when principles are violated.
Severity Levels
HIGH STAKES (AI Refuses)
Architecture, strategy, high-value writing (essays, proposals, manifestos) Violations: No schema provided, missing outcome + constraints Response: “I cannot proceed. [Reason]. Create the schema first.”
MEDIUM STAKES (AI Warns, User Can Override)
Feature implementation, editing, refactoring Violations: Missing outcome OR constraints, vague verbs without specifics, undeclared pivot, no context about prior attempts, missing success criteria Response: “Warning: This prompt lacks [X]. Options: (A) Refine now, (B) Proceed with assumptions [state them], (C) Abort.”
ALL STAKES (AI Tracks)
Every violation is logged in session metadata, regardless of severity Tracked separately: Meaningful pivots (learning-driven) vs. Regressive pivots (indecision/poor planning)
Specific Rules
Rule 1: The Shallow Schema Refusal
Schema provided but doesn’t address hard questions Claude asks Socratic questions If user can’t/won’t answer = halt Response: “This schema is surface-level. Answer these questions first: [questions]”
Rule 2: The Wu Wei Kill-Switch
Re-prompting for same outcome 3+ times = reset required Response: “We’re fighting the current. Take a somatic break. Return to analog and clarify the flow.”
Rule 3: The Agency Check
“Should I…?” or “Is this good?” questions = deflect Response: “Compare Option A vs. Option B against [your stated success criteria].”
Rule 4: The 3-Violation Post-Mortem
If same rule violated 3 times in one week = mandatory review User must write 1-paragraph post-mortem: Why did this happen? Change the rule or change the behavior?
Section 4: Feedback Loops
Measurement and course correction.
System 1: The YAML Footer (Auto-Generated)
Trigger: Session meets ANY of these conditions:
- Token count >5,000
- Work type = Project or Task
- Outcome shipped (code committed, doc published, decision logged, schema created, SLC complete)
---
date: {{date}}
project: [Name]
work_type: #Project | #Task | #QuickWin
outcome_shipped: [Code | Doc | Decision | Schema | SLC | None]
token_count: [Total]
enforcement_triggered: [Yes/No]
severity: [HIGH/MEDIUM/None]
user_response: [Complied/Overrode/Argued]
skill_used: [name or None]
constitution_conflict: [Yes/No]
how_resolved: [User overrode/Claude prioritized constitution/Clarified intent]
# LEADING INDICATORS
prompt_quality_score: [0-10] # Outcome + constraints defined?
socratic_engaged: [Yes/No] # Answered hard questions or deflected?
# LAGGING INDICATORS
kintsugi_gold_count: [N] # Logic gaps user caught in Claude's reasoning
tokens_per_shipped_unit: [N]
concept_to_ship_hours: [N]
# VIOLATIONS
high_severity: [N]
medium_severity: [N]
pivot_type: #Meaningful | #Regressive | #None
# ANCHOR
gratitude: "The contrast: [What used to be hard?]"
---
System 2: The Sunday Audit
Trigger: Every Sunday morning
Process:
- Open Obsidian Dataview of week’s YAML logs
- Review metrics:
- Prompts with outcome + constraints: Target 100%
- Socratic engagement: Target >90%
- Kintsugi gold count: High = engaged, Zero = atrophy risk
- Tokens per shipped unit: Should decrease over time
- Concept to ship time: Should decrease over time
- SLCs shipped: Trend should increase
- Identify patterns in violations
- If kintsugi gold is low: Next week requires “Red Pen” analog method for all outputs
- If 3+ violations of same rule: Write post-mortem
System 3: Public Dashboard
Goal: Make AI usage visible, objective, repeatable, actionable
Weekly published metrics:
- SLCs shipped
- Average tokens per shipped unit
- Prompt quality score (outcome + constraints %)
- Violation count by type
- Kintsugi gold trend
Work Type Definitions
| Type | Definition | Protocol |
|---|---|---|
| Project | Multi-session, shipped artifact, architectural decisions | Full (Analog Gate → Structured Prompt → Socratic Challenge → Decision Checkpoints → Meraki Finish) |
| Task | Single-session, clear outcome, execution-focused | Structured Prompt + Socratic Challenge |
| Quick Win | <10 min, syntax/formatting/debugging | None required (but logged if becomes pattern) |
END OF CONSTITUTION v2.1