tractatus/docs/markdown/tractatus-ai-safety-framework-core-values-and-principles.md
TheFlow 2298d36bed fix(submissions): restructure Economist package and fix article display
- Create Economist SubmissionTracking package correctly:
  * mainArticle = full blog post content
  * coverLetter = 216-word SIR— letter
  * Links to blog post via blogPostId
- Archive 'Letter to The Economist' from blog posts (it's the cover letter)
- Fix date display on article cards (use published_at)
- Target publication already displaying via blue badge

Database changes:
- Make blogPostId optional in SubmissionTracking model
- Economist package ID: 68fa85ae49d4900e7f2ecd83
- Le Monde package ID: 68fa2abd2e6acd5691932150

Next: Enhanced modal with tabs, validation, export

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-24 08:47:42 +13:00

23 KiB

Tractatus AI Safety Framework - Core Values and Principles

Document Type: Strategic Foundation Created: 2025-10-06 Author: John Stroh Version: 1.0 Status: Active


Purpose

This document establishes the foundational values and principles that guide the Tractatus AI Safety Framework and all aspects of this website platform. These enduring elements represent our deepest commitments to safe AI development and provide the basis for strategic alignment across all features, content, and operations.


Core Values

Sovereignty & Self-determination

  • Human Agency Preservation: AI systems must augment, never replace, human decision-making authority
  • User Control: Individuals maintain complete control over their data and engagement with AI features
  • No Manipulation: Zero dark patterns, no hidden AI influence, complete transparency in AI operations
  • Explicit Consent: All AI features require clear user understanding and opt-in

Transparency & Honesty

  • Visible AI Reasoning: All AI-generated suggestions include the reasoning process
  • Public Moderation Queue: Human oversight decisions are documented and visible
  • Clear Boundaries: Explicitly communicate what AI can and cannot do
  • Honest Limitations: Acknowledge framework limitations and edge cases
  • No Proprietary Lock-in: Open source, open standards, exportable data

Harmlessness & Protection

  • Privacy-First Design: No tracking, no surveillance, minimal data collection
  • Security by Default: Regular audits, penetration testing, zero-trust architecture
  • Fail-Safe Mechanisms: AI errors default to human review, not automatic action
  • Boundary Enforcement: Architectural design prevents AI from making values decisions
  • User Safety: Protection from AI-generated misinformation or harmful content

Human Judgment Primacy

  • Values Decisions: Always require human approval, never delegated to AI
  • Strategic Oversight: Human authority over mission, values, and governance
  • Escalation Protocols: Clear pathways for AI to request human guidance
  • Override Capability: Humans can always override AI suggestions
  • Accountability: Human responsibility for all AI-assisted actions

Community & Accessibility

  • Universal Access: Core framework documentation freely available to all
  • Three Audience Paths: Tailored content for Researchers, Implementers, Advocates
  • Economic Accessibility: Free tier with substantive capabilities
  • Knowledge Sharing: Open collaboration, peer review, community contributions
  • WCAG Compliance: Accessible to all abilities and assistive technologies

Biodiversity & Ecosystem Thinking

  • Multiple Valid Approaches: No single solution, respect for alternative frameworks
  • Interoperability: Integration with diverse AI safety approaches
  • Sustainability: Long-term viability over short-term growth
  • Resilience: Distributed systems, multiple mirrors, no single points of failure
  • Environmental Responsibility: Green hosting, efficient code, minimal resource consumption

Guiding Principles

Architectural Safety Enforcement

  • Structural over Training: Safety through architecture, not just fine-tuning
  • Explicit Boundaries: Codified limits on AI action authority
  • Verifiable Compliance: Automated checks against strategic values
  • Cross-Reference Validation: AI actions validated against explicit instructions
  • Context Pressure Monitoring: Detection of error-prone conditions

Dogfooding Implementation

  • Self-Application: This website uses Tractatus to govern its own AI operations
  • Living Demonstration: Platform proves framework effectiveness through use
  • Continuous Validation: Real-world testing of governance mechanisms
  • Transparent Meta-Process: Public documentation of how AI governs AI

Progressive Implementation

  • Phased Rollout: 4-phase deployment over 18 months
  • Incremental Features: Add capabilities as governance matures
  • No Shortcuts: Quality over speed, world-class execution
  • Learn and Adapt: Iterate based on real-world feedback

Education-Centered Approach

  • Demystify AI Safety: Make complex concepts accessible
  • Build Literacy: Empower users to understand AI governance
  • Interactive Demonstrations: Learn by doing (classification, 27027 incident, boundary enforcement)
  • Case Study Learning: Real-world failures and successes
  • Open Research: Share findings, encourage replication

Jurisdictional Awareness & Data Sovereignty

  • Respect Indigenous Leadership: Honor indigenous data sovereignty principles (CARE Principles)
  • Te Tiriti Foundation: Acknowledge Te Tiriti o Waitangi as strategic baseline
  • Location-Aware Hosting: Consider data residency and jurisdiction
  • Global Application: Framework designed for worldwide implementation
  • Local Adaptation: Support for cultural and legal contexts

AI Governance Framework

  • Quadrant-Based Classification: Strategic/Operational/Tactical/System/Stochastic organization
  • Time-Persistence Metadata: Instructions classified by longevity and importance
  • Human-AI Collaboration: Clear delineation of authority and responsibility
  • Instruction Persistence: Critical directives maintained across context windows
  • Metacognitive Verification: AI self-assessment before proposing actions

Research & Validation Priority

  • Peer Review: Academic rigor, scholarly publication
  • Reproducible Results: Open code, documented methodologies
  • Falsifiability: Framework designed to be tested and potentially disproven
  • Continuous Research: Ongoing validation and refinement
  • Industry Collaboration: Partnerships with AI organizations

Sustainable Operations

  • Koha Model: Transparent, community-supported funding (Phase 3+)
  • No Exploitation: Fair pricing, clear value exchange
  • Resource Efficiency: Optimized code, cached content, minimal overhead
  • Long-Term Thinking: Decades, not quarters
  • Community Ownership: Contributors have stake in success

Te Tiriti o Waitangi Commitment

Strategic Baseline (Not Dominant Cultural Overlay):

The Tractatus framework acknowledges Te Tiriti o Waitangi and indigenous leadership in digital sovereignty as a strategic foundation for this work. We:

  • Respect Indigenous Data Sovereignty: Follow documented principles (CARE Principles, Te Mana Raraunga research)
  • Acknowledge Historical Leadership: Indigenous peoples have led sovereignty struggles for centuries
  • Apply Published Standards: Use peer-reviewed indigenous data governance frameworks
  • Defer Deep Engagement: Will wait to approach Māori organizations until we have a stable and well developed platform in production. Our objective will be to request help in editing a Māori version that has their support and approval.

Implementation:

  • Footer acknowledgment (subtle, respectful)
  • /about/values page (detailed explanation)
  • Resource directory (links to Māori data sovereignty work)
  • No tokenism, no performative gestures

Values Alignment in Practice

Content Curation (Blog, Resources)

  • AI Suggests: Claude analyzes trends, proposes topics
  • Human Approves: All values-sensitive content requires human review
  • Transparency: AI reasoning visible in moderation queue
  • Attribution: Clear "AI-curated, human-approved" labels

Media Inquiries

  • AI Triages: Analyzes urgency, topic sensitivity
  • Human Responds: All responses written or approved by humans
  • Escalation: Values-sensitive topics immediately escalated to strategic review

Case Study Submissions

  • AI Reviews: Assesses relevance, completeness
  • Human Validates: Final publication decision always human
  • Quality Control: Framework alignment checked against TRA-VAL-0001

Interactive Demonstrations

  • Educational Purpose: Teach framework concepts through interaction
  • No Live Data: Demonstrations use example scenarios only
  • Transparency: Show exactly how classification and validation work

Decision Framework

When values conflict (e.g., transparency vs. privacy, speed vs. safety):

  1. Explicit Recognition: Acknowledge the tension publicly
  2. Context Analysis: Consider specific situation and stakeholders
  3. Hierarchy Application:
    • Human Safety > System Performance
    • Privacy > Convenience
    • Transparency > Proprietary Advantage
    • Long-term Sustainability > Short-term Growth
  4. Document Resolution: Record decision rationale for future reference
  5. Community Input: Seek feedback on significant value trade-offs

Review and Evolution

Annual Review Process

  • Scheduled: 2026-10-06 (one year from creation)
  • Scope: Comprehensive evaluation of values relevance and implementation
  • Authority: Human PM (John Stroh) with community input
  • Outcome: Updated version or reaffirmation of current values

Triggering Extraordinary Review

Immediate review required if:

  • Framework fails to prevent significant AI harm
  • Values found to be in conflict with actual operations
  • Major regulatory or ethical landscape changes
  • Community identifies fundamental misalignment

Evolution Constraints

  • Core values (Sovereignty, Transparency, Harmlessness, Human Judgment) are immutable
  • Guiding principles may evolve based on evidence and experience
  • Changes require explicit human approval and public documentation

Metrics for Values Adherence

Sovereignty & Self-determination

  • Zero instances of hidden AI influence
  • 100% opt-in for AI features
  • User data export capability maintained

Transparency & Honesty

  • All AI reasoning documented in moderation queue
  • Public disclosure of framework limitations
  • Clear attribution of AI vs. human content

Harmlessness & Protection

  • Zero security breaches
  • Privacy audit pass rate: 100%
  • Fail-safe activation rate (AI defers to human)

Human Judgment Primacy

  • 100% of values decisions reviewed by humans
  • Average escalation response time < 24 hours
  • Zero unauthorized AI autonomous actions

Community & Accessibility

  • WCAG AA compliance: 100% of pages
  • Free tier usage: >80% of all users
  • Community contributions accepted and integrated

Implementation Requirements

All features, content, and operations must:

  1. Pass Values Alignment Check: Documented review against this framework
  2. Include Tractatus Governance: Boundary enforcement, classification, validation
  3. Maintain Human Oversight: Clear escalation paths to human authority
  4. Support Transparency: Reasoning and decision processes visible
  5. Respect User Sovereignty: No manipulation, complete control, clear consent

Failure to align with these values is grounds for feature rejection or removal.


Appendix A: Values in Action Examples

Example 1: Blog Post Suggestion

AI Action: Suggests topic "Is AI Safety Overblown?" Classification: STOCHASTIC (exploration) → escalate to STRATEGIC (values-sensitive) Human Review: Topic involves framework credibility, requires strategic approval Decision: Approved with requirement for balanced, evidence-based treatment Outcome: Blog post published with AI reasoning visible, cites peer-reviewed research

Example 2: Media Inquiry Response

AI Action: Triages inquiry from major tech publication as "high urgency" Classification: OPERATIONAL (standard process) Human Review: Response drafted by human, reviews AI summary for accuracy Decision: Human-written response sent, AI triage saved time Outcome: Effective media engagement, human authority maintained

Example 3: Feature Request

AI Action: Suggests adding "auto-approve" for low-risk blog posts Classification: STRATEGIC (changes governance boundary) Human Review: Would reduce human oversight, conflicts with core values Decision: Rejected - all content requires human approval per TRA-VAL-0001 Outcome: Framework integrity preserved, alternative efficiency improvements explored


Appendix B: Glossary

AI Governance: Frameworks and mechanisms that control AI system behavior Boundary Enforcement: Preventing AI from actions outside defined authority Dogfooding: Using the framework to govern itself (meta-implementation) Human Judgment Primacy: Core principle that humans retain decision authority Quadrant Classification: Strategic/Operational/Tactical/System/Stochastic categorization Time-Persistence Metadata: Instruction classification by longevity and importance Values-Sensitive: Content or decisions that intersect with strategic values


Document Metadata

  • Version: 1.0
  • Created: 2025-10-06
  • Last Modified: 2025-10-13
  • Author: John Stroh
  • Word Count: 1,717 words
  • Reading Time: ~9 minutes
  • Document ID: tractatus-ai-safety-framework-core-values-and-principles
  • Status: Active

License

Copyright 2025 John Stroh

Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at:

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.

Full License Text:

Apache License, Version 2.0, January 2004 http://www.apache.org/licenses/

TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION

  1. Definitions.

"License" shall mean the terms and conditions for use, reproduction, and distribution as defined by Sections 1 through 9 of this document.

"Licensor" shall mean the copyright owner or entity authorized by the copyright owner that is granting the License.

"Legal Entity" shall mean the union of the acting entity and all other entities that control, are controlled by, or are under common control with that entity. For the purposes of this definition, "control" means (i) the power, direct or indirect, to cause the direction or management of such entity, whether by contract or otherwise, or (ii) ownership of fifty percent (50%) or more of the outstanding shares, or (iii) beneficial ownership of such entity.

"You" (or "Your") shall mean an individual or Legal Entity exercising permissions granted by this License.

"Source" form shall mean the preferred form for making modifications, including but not limited to software source code, documentation source, and configuration files.

"Object" form shall mean any form resulting from mechanical transformation or translation of a Source form, including but not limited to compiled object code, generated documentation, and conversions to other media types.

"Work" shall mean the work of authorship, whether in Source or Object form, made available under the License, as indicated by a copyright notice that is included in or attached to the work.

"Derivative Works" shall mean any work, whether in Source or Object form, that is based on (or derived from) the Work and for which the editorial revisions, annotations, elaborations, or other modifications represent, as a whole, an original work of authorship. For the purposes of this License, Derivative Works shall not include works that remain separable from, or merely link (or bind by name) to the interfaces of, the Work and Derivative Works thereof.

"Contribution" shall mean any work of authorship, including the original version of the Work and any modifications or additions to that Work or Derivative Works thereof, that is intentionally submitted to Licensor for inclusion in the Work by the copyright owner or by an individual or Legal Entity authorized to submit on behalf of the copyright owner. For the purposes of this definition, "submitted" means any form of electronic, verbal, or written communication sent to the Licensor or its representatives, including but not limited to communication on electronic mailing lists, source code control systems, and issue tracking systems that are managed by, or on behalf of, the Licensor for the purpose of discussing and improving the Work, but excluding communication that is conspicuously marked or otherwise designated in writing by the copyright owner as "Not a Contribution."

"Contributor" shall mean Licensor and any individual or Legal Entity on behalf of whom a Contribution has been received by Licensor and subsequently incorporated within the Work.

  1. Grant of Copyright License. Subject to the terms and conditions of this License, each Contributor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable copyright license to reproduce, prepare Derivative Works of, publicly display, publicly perform, sublicense, and distribute the Work and such Derivative Works in Source or Object form.

  2. Grant of Patent License. Subject to the terms and conditions of this License, each Contributor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable (except as stated in this section) patent license to make, have made, use, offer to sell, sell, import, and otherwise transfer the Work, where such license applies only to those patent claims licensable by such Contributor that are necessarily infringed by their Contribution(s) alone or by combination of their Contribution(s) with the Work to which such Contribution(s) was submitted. If You institute patent litigation against any entity (including a cross-claim or counterclaim in a lawsuit) alleging that the Work or a Contribution incorporated within the Work constitutes direct or contributory patent infringement, then any patent licenses granted to You under this License for that Work shall terminate as of the date such litigation is filed.

  3. Redistribution. You may reproduce and distribute copies of the Work or Derivative Works thereof in any medium, with or without modifications, and in Source or Object form, provided that You meet the following conditions:

    (a) You must give any other recipients of the Work or Derivative Works a copy of this License; and

    (b) You must cause any modified files to carry prominent notices stating that You changed the files; and

    (c) You must retain, in the Source form of any Derivative Works that You distribute, all copyright, patent, trademark, and attribution notices from the Source form of the Work, excluding those notices that do not pertain to any part of the Derivative Works; and

    (d) If the Work includes a "NOTICE" text file as part of its distribution, then any Derivative Works that You distribute must include a readable copy of the attribution notices contained within such NOTICE file, excluding those notices that do not pertain to any part of the Derivative Works, in at least one of the following places: within a NOTICE text file distributed as part of the Derivative Works; within the Source form or documentation, if provided along with the Derivative Works; or, within a display generated by the Derivative Works, if and wherever such third-party notices normally appear. The contents of the NOTICE file are for informational purposes only and do not modify the License. You may add Your own attribution notices within Derivative Works that You distribute, alongside or as an addendum to the NOTICE text from the Work, provided that such additional attribution notices cannot be construed as modifying the License.

    You may add Your own copyright statement to Your modifications and may provide additional or different license terms and conditions for use, reproduction, or distribution of Your modifications, or for any such Derivative Works as a whole, provided Your use, reproduction, and distribution of the Work otherwise complies with the conditions stated in this License.

  4. Submission of Contributions. Unless You explicitly state otherwise, any Contribution intentionally submitted for inclusion in the Work by You to the Licensor shall be under the terms and conditions of this License, without any additional terms or conditions. Notwithstanding the above, nothing herein shall supersede or modify the terms of any separate license agreement you may have executed with Licensor regarding such Contributions.

  5. Trademarks. This License does not grant permission to use the trade names, trademarks, service marks, or product names of the Licensor, except as required for reasonable and customary use in describing the origin of the Work and reproducing the content of the NOTICE file.

  6. Disclaimer of Warranty. Unless required by applicable law or agreed to in writing, Licensor provides the Work (and each Contributor provides its Contributions) on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied, including, without limitation, any warranties or conditions of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A PARTICULAR PURPOSE. You are solely responsible for determining the appropriateness of using or redistributing the Work and assume any risks associated with Your exercise of permissions under this License.

  7. Limitation of Liability. In no event and under no legal theory, whether in tort (including negligence), contract, or otherwise, unless required by applicable law (such as deliberate and grossly negligent acts) or agreed to in writing, shall any Contributor be liable to You for damages, including any direct, indirect, special, incidental, or consequential damages of any character arising as a result of this License or out of the use or inability to use the Work (including but not limited to damages for loss of goodwill, work stoppage, computer failure or malfunction, or any and all other commercial damages or losses), even if such Contributor has been advised of the possibility of such damages.

  8. Accepting Warranty or Additional Liability. While redistributing the Work or Derivative Works thereof, You may choose to offer, and charge a fee for, acceptance of support, warranty, indemnity, or other liability obligations and/or rights consistent with this License. However, in accepting such obligations, You may act only on Your own behalf and on Your sole responsibility, not on behalf of any other Contributor, and only if You agree to indemnify, defend, and hold each Contributor harmless for any liability incurred by, or claims asserted against, such Contributor by reason of your accepting any such warranty or additional liability.

END OF TERMS AND CONDITIONS


Document Authority: This document has final authority over all platform operations. In case of conflict between this document and any other guidance, TRA-VAL-0001 takes precedence.

Next Review: 2026-10-06 Version History: v1.0 (2025-10-06) - Initial creation


This document is maintained by John Stroh (john.stroh.nz@pm.me) and subject to annual review. Changes require explicit human approval and public documentation.