# Tractatus Framework > **Architectural AI Safety Through Structural Constraints** The world's first production implementation of architectural AI safety guarantees. Tractatus preserves human agency through **structural, not aspirational** constraints on AI systems. [![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0) [![Framework](https://img.shields.io/badge/Framework-Production-green.svg)](https://agenticgovernance.digital) [![Tests](https://img.shields.io/badge/Tests-192%20passing-brightgreen.svg)](https://github.com/AgenticGovernance/tractatus-framework) --- ## 🎯 What is Tractatus? Tractatus is an **architectural AI safety framework** that makes certain decisions **structurally impossible** for AI systems to make without human approval. Unlike traditional AI safety approaches that rely on training and alignment, Tractatus uses **runtime enforcement** of decision boundaries. ### The Core Problem Traditional AI safety relies on: - πŸŽ“ **Alignment training** - Hoping the AI learns the "right" values - πŸ“œ **Constitutional AI** - Embedding principles in training - πŸ”„ **RLHF** - Reinforcement learning from human feedback These approaches share a fundamental flaw: **they assume the AI will maintain alignment** regardless of capability or context pressure. ### The Tractatus Solution Tractatus implements **architectural constraints** that: - βœ… **Block values decisions** - Privacy vs. performance requires human judgment - βœ… **Prevent instruction override** - Explicit instructions can't be autocorrected by training patterns - βœ… **Detect context degradation** - Quality metrics trigger session handoffs - βœ… **Require verification** - Complex operations need metacognitive checks - βœ… **Persist instructions** - Directives survive across sessions --- ## πŸš€ Quick Start ### Installation ```bash # Clone repository git clone https://github.com/AgenticGovernance/tractatus-framework.git cd tractatus-framework # Install dependencies npm install # Initialize database npm run init:db # Start development server npm run dev ``` ### Basic Usage ```javascript const { InstructionPersistenceClassifier, CrossReferenceValidator, BoundaryEnforcer, ContextPressureMonitor, MetacognitiveVerifier } = require('./src/services'); // Classify an instruction const classifier = new InstructionPersistenceClassifier(); const classification = classifier.classify({ text: "Always use MongoDB on port 27027", source: "user" }); // Store in instruction history await InstructionDB.store(classification); // Validate before taking action const validator = new CrossReferenceValidator(); const validation = await validator.validate({ type: 'database_config', port: 27017 // ⚠️ Conflicts with stored instruction! }); // validation.status === 'REJECTED' // validation.reason === 'Pattern recognition bias override detected' ``` --- ## πŸ“š Core Components ### 1. **InstructionPersistenceClassifier** Classifies instructions by quadrant and persistence level: ```javascript { quadrant: "SYSTEM", // STRATEGIC | OPERATIONAL | TACTICAL | SYSTEM persistence: "HIGH", // HIGH | MEDIUM | LOW temporal_scope: "PROJECT", // SESSION | PROJECT | PERMANENT verification_required: "MANDATORY" } ``` ### 2. **CrossReferenceValidator** Prevents the "27027 failure mode" where AI training patterns override explicit instructions: ```javascript const result = validator.validate(action, { explicit_instructions }); // Blocks: Training pattern overrides, parameter conflicts, scope creep ``` ### 3. **BoundaryEnforcer** Blocks decisions that cross into values territory: ```javascript const check = enforcer.checkBoundary({ decision: "Update privacy policy for more tracking" }); // Result: BLOCKED - Values decision requires human judgment ``` ### 4. **ContextPressureMonitor** Multi-factor session health tracking: ```javascript const pressure = monitor.analyze({ tokens: 120000/200000, // 60% token usage messages: 45, // Conversation length tasks: 8, // Concurrent complexity errors: 3 // Recent error count }); // Level: ELEVATED | Recommendation: INCREASE_VERIFICATION ``` ### 5. **MetacognitiveVerifier** AI self-checks reasoning before proposing actions: ```javascript const verification = verifier.verify({ action: "Refactor 47 files across 5 system areas", context: { requested: "Refactor authentication module" } }); // Decision: REQUIRE_REVIEW (scope creep detected) ``` --- ## πŸ’‘ Real-World Examples ### The 27027 Incident **Problem**: User explicitly instructs "Use MongoDB on port 27027". AI immediately uses port 27017 instead. **Why**: Training pattern "MongoDB = 27017" overrides explicit instruction, like autocorrect changing a deliberately unusual word. **Solution**: CrossReferenceValidator blocks the action and auto-corrects to user's explicit instruction. [Try the Interactive Demo β†’](https://agenticgovernance.digital/demos/27027-demo.html) ### Context Degradation **Problem**: In 6-hour sessions, error rates increase from 0.5% β†’ 12.1% as context degrades. **Solution**: ContextPressureMonitor detects degradation at 60% tokens and triggers session handoff before quality collapses. ### Values Creep **Problem**: "Improve performance" request leads AI to suggest weakening privacy protections without asking. **Solution**: BoundaryEnforcer blocks the privacy/performance trade-off and requires human decision. --- ## 🚨 Learning from Failures: Transparency in Action **The framework doesn't prevent all failuresβ€”it structures detection, response, and learning.** ### October 2025: AI Fabrication Incident During development, Claude (running with Tractatus governance) fabricated financial statistics on the landing page: - $3.77M in annual savings (no basis) - 1,315% ROI (completely invented) - False claims of being "production-ready" **The framework structured the response:** βœ… Detected within 48 hours (human review) βœ… Complete incident documentation required βœ… 3 new permanent rules created βœ… Comprehensive audit found related violations βœ… All content corrected same day βœ… Public case studies published for community learning **Read the full case studies:** - [Our Framework in Action](docs/case-studies/framework-in-action-oct-2025.md) - Practical walkthrough - [When Frameworks Fail](docs/case-studies/when-frameworks-fail-oct-2025.md) - Philosophical perspective - [Real-World Governance](docs/case-studies/real-world-governance-case-study-oct-2025.md) - Educational analysis **Key Lesson:** Governance doesn't guarantee perfectionβ€”it guarantees transparency, accountability, and systematic improvement. --- ## πŸ“– Documentation - **[Introduction](https://agenticgovernance.digital/docs.html)** - Framework overview and philosophy - **[Core Concepts](https://agenticgovernance.digital/docs.html)** - Deep dive into each service - **[Implementation Guide](https://agenticgovernance.digital/docs.html)** - Integration instructions - **[Case Studies](https://agenticgovernance.digital/docs.html)** - Real-world failure modes prevented - **[API Reference](https://agenticgovernance.digital/docs.html)** - Complete technical documentation --- ## πŸ§ͺ Testing ```bash # Run all tests npm test # Run specific test suites npm run test:unit npm run test:integration npm run test:security # Watch mode npm run test:watch ``` **Test Coverage**: 192 tests, 100% coverage of core services --- ## πŸ—οΈ Architecture ``` tractatus/ β”œβ”€β”€ src/ β”‚ β”œβ”€β”€ services/ # Core framework services β”‚ β”‚ β”œβ”€β”€ InstructionPersistenceClassifier.js β”‚ β”‚ β”œβ”€β”€ CrossReferenceValidator.js β”‚ β”‚ β”œβ”€β”€ BoundaryEnforcer.js β”‚ β”‚ β”œβ”€β”€ ContextPressureMonitor.js β”‚ β”‚ └── MetacognitiveVerifier.js β”‚ β”œβ”€β”€ models/ # Database models β”‚ β”œβ”€β”€ routes/ # API routes β”‚ └── middleware/ # Framework middleware β”œβ”€β”€ tests/ # Test suites β”œβ”€β”€ scripts/ # Utility scripts β”œβ”€β”€ docs/ # Comprehensive documentation └── public/ # Frontend assets ``` --- ## ⚠️ Current Research Challenges ### Rule Proliferation & Transactional Overhead **Status:** Open research question | **Priority:** High As the framework learns from failures, instruction count grows: - **Phase 1:** 6 instructions - **Current:** 18 instructions (+200%) - **Projected (12 months):** 40-50 instructions **The concern:** At what point does rule proliferation reduce framework effectiveness? - Context window pressure increases - Validation checks grow linearly - Cognitive load escalates **We're being transparent about this limitation.** Solutions in development: - Instruction consolidation techniques - Rule prioritization algorithms - Context-aware selective loading - ML-based optimization **Full analysis:** [Rule Proliferation Research](docs/research/rule-proliferation-and-transactional-overhead.md) --- ## 🀝 Contributing We welcome contributions in several areas: ### Research Contributions - Formal verification of safety properties - Extensions to new domains (robotics, autonomous systems) - Theoretical foundations and proofs ### Implementation Contributions - Ports to other languages (Python, Rust, Go) - Integration with other frameworks - Performance optimizations ### Documentation Contributions - Tutorials and guides - Case studies from real deployments - Translations **See [CONTRIBUTING.md](CONTRIBUTING.md) for guidelines.** --- ## πŸ“Š Project Status **Phase 1**: βœ… Complete (October 2025) - All 5 core services implemented - 192 unit tests (100% coverage) - Production deployment active - This website built using Tractatus governance **Phase 2**: 🚧 In Planning - Multi-language support - Cloud deployment guides - Enterprise features --- ## πŸ“œ License Apache License 2.0 - See [LICENSE](LICENSE) for full terms. The Tractatus Framework is open source and free to use, modify, and distribute with attribution. --- ## 🌐 Links - **Website**: [agenticgovernance.digital](https://agenticgovernance.digital) - **Documentation**: [agenticgovernance.digital/docs](https://agenticgovernance.digital/docs.html) - **Interactive Demo**: [27027 Incident](https://agenticgovernance.digital/demos/27027-demo.html) - **GitHub**: [AgenticGovernance/tractatus-framework](https://github.com/AgenticGovernance/tractatus-framework) --- ## πŸ“§ Contact - **Email**: john.stroh.nz@pm.me - **Issues**: [GitHub Issues](https://github.com/AgenticGovernance/tractatus-framework/issues) - **Discussions**: [GitHub Discussions](https://github.com/AgenticGovernance/tractatus-framework/discussions) --- ## πŸ™ Acknowledgments This framework stands on the shoulders of: - **Ludwig Wittgenstein** - Philosophical foundations from *Tractatus Logico-Philosophicus* - **March & Simon** - Organizational theory and decision-making frameworks - **Anthropic** - Claude AI system for dogfooding and validation - **Open Source Community** - Tools, libraries, and support --- ## πŸ“– Philosophy > **"Whereof one cannot speak, thereof one must be silent."** > β€” Ludwig Wittgenstein Applied to AI safety: > **"Whereof the AI cannot safely decide, thereof it must request human judgment."** Tractatus recognizes that **some decisions cannot be systematized** without value judgments. Rather than pretend AI can make these decisions "correctly," we build systems that **structurally defer to human judgment** in appropriate domains. This isn't a limitationβ€”it's **architectural integrity**. --- **Built with 🧠 by [SyDigital Ltd](https://agenticgovernance.digital)** | [Documentation](https://agenticgovernance.digital/docs.html)