Added complete architectural_principles section to all three language files: - English (en/values.json): baseline translations - German (de/values.json): complete German translations - French (fr/values.json): complete French translations Fixes rendering issue on values.html where raw i18n keys were displaying instead of translated content. Section includes: - Five Alexander principles (Deep Interlock, Structure-Preserving, Gradients, Living Process, Not-Separateness) - Each with title, description, and connection to core values - Monitoring note about October 2025 integration - TOC entry for navigation Technical terms preserved in English across all translations per i18n guidelines. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
159 lines
13 KiB
JSON
159 lines
13 KiB
JSON
{
|
||
"page": {
|
||
"title": "Values & Principles | Tractatus AI Safety Framework",
|
||
"description": "Our core values: sovereignty, transparency, harmlessness, and community. Including our commitment to Te Tiriti o Waitangi and indigenous data sovereignty."
|
||
},
|
||
"header": {
|
||
"title": "Values & Principles",
|
||
"subtitle": "The foundational values that guide the Tractatus Framework's development, governance, and community."
|
||
},
|
||
"toc": {
|
||
"heading": "Contents",
|
||
"core_values": "Core Values",
|
||
"sovereignty": "Sovereignty",
|
||
"transparency": "Transparency",
|
||
"harmlessness": "Harmlessness",
|
||
"community": "Community",
|
||
"architectural_principles": "Architectural Principles",
|
||
"te_tiriti": "Te Tiriti o Waitangi & Digital Sovereignty",
|
||
"indigenous": "Indigenous Data Sovereignty",
|
||
"governance": "Governance & Accountability"
|
||
},
|
||
"core_values": {
|
||
"heading": "Core Values",
|
||
"intro": "These four values form the foundation of the Tractatus Framework. They are not aspirational—they are architectural. The framework is designed to enforce these values through structure, not training."
|
||
},
|
||
"sovereignty": {
|
||
"title": "1. Sovereignty",
|
||
"principle": "Principle:",
|
||
"principle_text": "Individuals and communities must maintain control over decisions affecting their data, privacy, values, and agency. AI systems must preserve human sovereignty, not erode it.",
|
||
"practice_heading": "What This Means in Practice:",
|
||
"framework_heading": "Framework Implementation:"
|
||
},
|
||
"transparency": {
|
||
"title": "2. Transparency",
|
||
"principle": "Principle:",
|
||
"principle_text": "All AI decisions must be explainable, auditable, and reversible. No black boxes. Users deserve to understand how and why systems make choices.",
|
||
"practice_heading": "What This Means in Practice:",
|
||
"framework_heading": "Framework Implementation:"
|
||
},
|
||
"harmlessness": {
|
||
"title": "3. Harmlessness",
|
||
"principle": "Principle:",
|
||
"principle_text": "AI systems must not cause harm through action or inaction. This includes preventing drift, detecting degradation, and enforcing boundaries against values erosion.",
|
||
"practice_heading": "What This Means in Practice:",
|
||
"framework_heading": "Framework Implementation:"
|
||
},
|
||
"community": {
|
||
"title": "4. Community",
|
||
"principle": "Principle:",
|
||
"principle_text": "AI safety is a collective endeavor. We are committed to open collaboration, knowledge sharing, and empowering communities to shape the AI systems that affect their lives.",
|
||
"practice_heading": "What This Means in Practice:",
|
||
"framework_heading": "Framework Implementation:"
|
||
},
|
||
"architectural_principles": {
|
||
"heading": "Architectural Principles",
|
||
"intro": "Our values—sovereignty, transparency, harmlessness, community—guide <em>what</em> we build. But values alone don't prevent systems from drifting. We need architectural principles that show <em>how</em> to preserve values through structure, not aspiration.",
|
||
"alexander_credit": "Drawing on Christopher Alexander's work in architectural pattern languages, we've identified five principles that translate living systems thinking into governance architecture. These aren't metaphors—they're operational requirements that shape every framework decision.",
|
||
"deep_interlock": {
|
||
"title": "Deep Interlock",
|
||
"description": "Six governance services <strong>coordinate through mutual validation</strong>, not operate in silos. When BoundaryEnforcer detects a values conflict, CrossReferenceValidator checks if it aligns with stored instructions, ContextPressureMonitor assesses session conditions, and PluralisticDeliberationOrchestrator coordinates stakeholder deliberation if needed.",
|
||
"connection": "<strong>Connects to Transparency:</strong> Service coordination creates audit trails showing how governance decisions emerge from multiple reinforcing checks, not single-point failures."
|
||
},
|
||
"structure_preserving": {
|
||
"title": "Structure-Preserving",
|
||
"description": "Framework changes <strong>enhance without breaking</strong>. When we add new governance rules or refine service logic, historical audit logs remain interpretable. A decision made under framework v4.2 can still be understood in v4.4—institutional memory preserved across evolution.",
|
||
"connection": "<strong>Connects to Accountability:</strong> Structure-preserving transformations mean governance continuity. Organizations can demonstrate regulatory compliance across framework versions because the audit trail remains coherent."
|
||
},
|
||
"gradients": {
|
||
"title": "Gradients Not Binary",
|
||
"description": "Governance operates on <strong>intensity scales</strong> (NORMAL/ELEVATED/HIGH/CRITICAL), not yes/no switches. Context pressure monitoring adjusts behavior gradually as session conditions change—token usage climbs, message length increases, task complexity escalates. Nuanced response to risk, not mechanical on/off.",
|
||
"connection": "<strong>Connects to Harmlessness:</strong> Gradients prevent both under-response (missing risks) and over-response (alert fatigue). The system adapts governance intensity to match actual risk levels, like living systems responding to environmental stress."
|
||
},
|
||
"living_process": {
|
||
"title": "Living Process",
|
||
"description": "The framework <strong>evolves from operational failures</strong>, not predetermined roadmaps. When the \"27027 incident\" revealed pattern recognition bias (AI autocorrected user's explicit \"port 27027\" to \"port 27017\"), we didn't just document the failure—we built CrossReferenceValidator to prevent that class of error architecturally.",
|
||
"connection": "<strong>Connects to Community:</strong> Living process means continuous learning from real-world use. The framework grows smarter through operational experience shared across the community, not consultant wisdom imposed from above."
|
||
},
|
||
"not_separateness": {
|
||
"title": "Not-Separateness",
|
||
"description": "Governance is <strong>woven into the deployment architecture</strong>, not bolted on as afterthought. PreToolUse hooks intercept actions before execution. Services run in the critical path. Bypasses require explicit <code>--no-verify</code> flags and are logged. Enforcement is structural, not voluntary.",
|
||
"connection": "<strong>Connects to Sovereignty:</strong> Not-separateness ensures AI cannot bypass governance to override human agency. The architecture makes it structurally difficult to erode boundaries, preserving decision-making authority where it belongs—with affected humans."
|
||
},
|
||
"monitoring_note": "<strong>Note:</strong> These principles were integrated into the framework in October 2025. We're monitoring their effectiveness through audit log analysis and operational metrics. This is active research—we're learning whether architectural principles from the built environment translate meaningfully to AI governance."
|
||
},
|
||
"te_tiriti_section": {
|
||
"title": "Te Tiriti o Waitangi & Digital Sovereignty",
|
||
"context_label": "Context:",
|
||
"context_text": "The Tractatus Framework is developed in Aotearoa New Zealand. We acknowledge Te Tiriti o Waitangi (the Treaty of Waitangi, 1840) as the founding document of this nation, and recognize the ongoing significance of tino rangatiratanga (self-determination) and kaitiakitanga (guardianship) in the digital realm.",
|
||
"context_para2": "This acknowledgment is not performative. Digital sovereignty—the principle that communities control their own data and technology—has deep roots in indigenous frameworks that predate Western tech by centuries.",
|
||
"why_matters_heading": "Why This Matters for AI Safety",
|
||
"why_matters_text": "Te Tiriti o Waitangi establishes principles of partnership, protection, and participation. These principles directly inform the Tractatus Framework's approach to digital sovereignty:",
|
||
"rangatiratanga_label": "Rangatiratanga (sovereignty):",
|
||
"rangatiratanga_text": "Communities must control decisions affecting their data and values",
|
||
"kaitiakitanga_label": "Kaitiakitanga (guardianship):",
|
||
"kaitiakitanga_text": "AI systems must be stewards, not exploiters, of data and knowledge",
|
||
"mana_label": "Mana (authority & dignity):",
|
||
"mana_text": "Technology must respect human dignity and cultural context",
|
||
"whanaungatanga_label": "Whanaungatanga (relationships):",
|
||
"whanaungatanga_text": "AI safety is collective, not individual—relationships matter",
|
||
"approach_heading": "Our Approach",
|
||
"approach_intro": "We do not claim to speak for Māori or indigenous communities. Instead, we:",
|
||
"approach_frameworks_label": "Follow established frameworks:",
|
||
"approach_frameworks_text": "We align with Te Mana Raraunga (Māori Data Sovereignty Network) and CARE Principles for Indigenous Data Governance",
|
||
"approach_respect_label": "Respect without tokenism:",
|
||
"approach_respect_text": "Te Tiriti forms part of our strategic foundation, not a superficial overlay",
|
||
"approach_engagement_label": "Avoid premature engagement:",
|
||
"approach_engagement_text": "We will not approach Māori organizations for endorsement until we have demonstrated value and impact",
|
||
"approach_learn_label": "Document and learn:",
|
||
"approach_learn_text": "We study indigenous data sovereignty principles and incorporate them architecturally",
|
||
"principles_heading": "Te Tiriti Principles in Practice",
|
||
"partnership_label": "Partnership:",
|
||
"partnership_text": "AI systems should be developed in partnership with affected communities, not imposed upon them.",
|
||
"protection_label": "Protection:",
|
||
"protection_text": "The framework protects against values erosion, ensuring cultural contexts are not overridden by AI assumptions.",
|
||
"participation_label": "Participation:",
|
||
"participation_text": "Communities maintain agency over AI decisions affecting their data and values."
|
||
},
|
||
"indigenous_data": {
|
||
"title": "Indigenous Data Sovereignty",
|
||
"intro": "Indigenous data sovereignty is the principle that indigenous peoples have the right to control the collection, ownership, and application of their own data. This goes beyond privacy—it's about self-determination in the digital age.",
|
||
"care_heading": "CARE Principles for Indigenous Data Governance",
|
||
"care_intro": "The Tractatus Framework aligns with the CARE Principles, developed by indigenous data governance experts:",
|
||
"collective_benefit_label": "Collective Benefit",
|
||
"collective_benefit_text": "Data ecosystems shall be designed and function in ways that enable Indigenous Peoples to derive benefit from the data.",
|
||
"authority_label": "Authority to Control",
|
||
"authority_text": "Indigenous Peoples' rights and interests in Indigenous data must be recognized and their authority to control such data be empowered.",
|
||
"responsibility_label": "Responsibility",
|
||
"responsibility_text": "Those working with Indigenous data have a responsibility to share how data are used to support Indigenous Peoples' self-determination and collective benefit.",
|
||
"ethics_label": "Ethics",
|
||
"ethics_text": "Indigenous Peoples' rights and wellbeing should be the primary concern at all stages of the data life cycle and across the data ecosystem.",
|
||
"resources_heading": "Resources & Further Reading",
|
||
"tmr_title": "Te Mana Raraunga – Māori Data Sovereignty Network →",
|
||
"tmr_description": "Leading network advancing Māori data sovereignty in Aotearoa.",
|
||
"care_title": "CARE Principles for Indigenous Data Governance →",
|
||
"care_description": "International framework for indigenous data rights."
|
||
},
|
||
"governance": {
|
||
"title": "Governance & Accountability",
|
||
"intro": "Values without enforcement are aspirations. The Tractatus Framework implements these values through architectural governance:",
|
||
"strategic_review_heading": "Strategic Review Protocol",
|
||
"strategic_review_text": "Quarterly reviews of framework alignment with stated values. Any drift from sovereignty, transparency, harmlessness, or community principles triggers mandatory correction.",
|
||
"values_alignment_heading": "Values Alignment Framework",
|
||
"values_alignment_text": "All major decisions (architectural changes, partnerships, licensing) must pass values alignment check. If a decision would compromise any core value, it is rejected.",
|
||
"human_oversight_heading": "Human Oversight Requirements",
|
||
"human_oversight_text": "AI-generated content (documentation, code examples, case studies) requires human approval before publication. No AI makes values decisions without human judgment.",
|
||
"community_accountability_heading": "Community Accountability",
|
||
"community_accountability_text": "Open source development means community oversight. If we fail to uphold these values, the community can fork, modify, or create alternatives. This is by design."
|
||
},
|
||
"commitment": {
|
||
"title": "Our Commitment",
|
||
"intro": "These values are not negotiable. They form the architectural foundation of the Tractatus Framework. We commit to:",
|
||
"item1": "Preserving human sovereignty over values decisions",
|
||
"item2": "Maintaining radical transparency in all framework operations",
|
||
"item3": "Preventing harm through structural constraints, not promises",
|
||
"item4": "Building and empowering community, not extracting from it",
|
||
"item5": "Respecting Te Tiriti o Waitangi and indigenous data sovereignty",
|
||
"closing": "When in doubt, we choose human agency over AI capability. Always."
|
||
}
|
||
}
|