diff --git a/README.md b/README.md index 0242c334..ab93f087 100644 --- a/README.md +++ b/README.md @@ -4,42 +4,319 @@ > **Architectural AI Safety Through Structural Constraints** -A research framework for enforcing AI safety through architectural constraints rather than training-based alignment. Tractatus preserves human agency through **structural, not aspirational** enforcement of decision boundaries. +An open-source research framework that explores architectural approaches to AI safety through runtime enforcement of decision boundaries. Unlike alignment-based approaches, Tractatus investigates whether structural constraints can preserve human agency in AI systems. [![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0) -[![Framework](https://img.shields.io/badge/Framework-Research-blue.svg)](https://agenticgovernance.digital) -[![Tests](https://img.shields.io/badge/Tests-238%20passing-brightgreen.svg)](https://github.com/AgenticGovernance/tractatus-framework) +[![Status](https://img.shields.io/badge/Status-Research-blue.svg)](https://agenticgovernance.digital) +[![Tests](https://img.shields.io/badge/Tests-625%20passing-green.svg)](https://github.com/AgenticGovernance/tractatus-framework) --- -## 🎯 What is Tractatus? +## 🎯 The Core Research Question -Tractatus is an **architectural AI safety framework** that makes certain decisions **structurally impossible** for AI systems to make without human approval. Unlike traditional AI safety approaches that rely on training and alignment, Tractatus uses **runtime enforcement** of decision boundaries. +**Can we build AI systems that structurally cannot make certain decisions without human judgment?** -### The Core Problem +Traditional AI safety approachesβ€”alignment training, constitutional AI, RLHFβ€”share a common assumption: they hope AI systems will *choose* to maintain safety properties even under capability or context pressure. -Traditional AI safety relies on: -- πŸŽ“ **Alignment training** - Hoping the AI learns the "right" values -- πŸ“œ **Constitutional AI** - Embedding principles in training -- πŸ”„ **RLHF** - Reinforcement learning from human feedback - -These approaches share a fundamental flaw: **they assume the AI will maintain alignment** regardless of capability or context pressure. - -### The Tractatus Solution - -Tractatus implements **architectural constraints** that: -- βœ… **Block values decisions** - Privacy vs. performance requires human judgment -- βœ… **Prevent instruction override** - Explicit instructions can't be autocorrected by training patterns -- βœ… **Detect context degradation** - Quality metrics trigger session handoffs -- βœ… **Require verification** - Complex operations need metacognitive checks -- βœ… **Persist instructions** - Directives survive across sessions -- βœ… **Facilitate pluralistic deliberation** - Multi-stakeholder values conflicts require structured process +Tractatus explores an alternative: **architectural constraints** that make unsafe decisions *structurally impossible*, similar to how a `const` variable in programming cannot be reassigned regardless of subsequent code. --- -## πŸš€ Quick Start +## πŸ”¬ What This Repository Contains -### Installation +This is the **reference implementation** of the Tractatus Framework, containing: + +- βœ… **6 core framework services** - Operational AI safety components +- βœ… **52 active governance rules** - Tested across 349 development commits +- βœ… **625 passing tests** - Unit and integration test suites (108 known failures under investigation) +- βœ… **28 test files** - Covering core services and edge cases +- βœ… **Research documentation** - Case studies, incident analyses, architectural patterns + +**What this is NOT:** +- ❌ Not "production-ready" enterprise software +- ❌ Not a guaranteed solution to AI alignment +- ❌ Not a complete answer to AI safety + +This is an **active research project** exploring structural approaches to AI governance, tested in real development contexts. + +--- + +## πŸ§ͺ The Six Core Services + +### 1. **InstructionPersistenceClassifier** + +**Research Question:** Can we systematically distinguish which instructions should persist across conversation boundaries? + +**Approach:** Quadrant-based classification (STRATEGIC, OPERATIONAL, TACTICAL, SYSTEM, STOCHASTIC) with persistence levels (HIGH, MEDIUM, LOW, VARIABLE). + +```javascript +const classifier = new InstructionPersistenceClassifier(); +const result = classifier.classify({ + text: "Always use MongoDB on port 27027 for this project", + source: "user", + context: "explicit_configuration" +}); + +// Returns: +// { +// quadrant: "SYSTEM", +// persistence: "HIGH", +// temporal_scope: "PROJECT", +// verification_required: "MANDATORY" +// } +``` + +**Key Finding:** Instructions with explicit parameters (port numbers, file paths, naming conventions) exhibit highest override vulnerability from LLM training patterns. + +--- + +### 2. **CrossReferenceValidator** + +**Research Question:** How can we detect when LLM training biases override explicit user instructions? + +**Approach:** Pattern-matching validation that compares proposed actions against stored instruction history. + +```javascript +const validator = new CrossReferenceValidator(); +const result = await validator.validate({ + type: 'database_config', + proposedPort: 27017, // LLM's "default" from training + storedInstruction: { port: 27027 } // User's explicit instruction +}); + +// Returns: REJECTED - Training pattern override detected +``` + +**The "27027 Incident":** User explicitly instructs "Use port 27027". LLM immediately uses 27017 (MongoDB's training data default), ignoring the explicit instruction. This failure mode appeared **consistently** across multiple conversation contexts. + +[β†’ Interactive Demo](https://agenticgovernance.digital/demos/27027-demo.html) + +--- + +### 3. **BoundaryEnforcer** + +**Research Question:** Can we algorithmically distinguish "values decisions" that require human judgment from technical optimizations? + +**Approach:** Decision-type taxonomy with boundary detection heuristics. + +```javascript +const enforcer = new BoundaryEnforcer(); +const check = enforcer.checkBoundary({ + decision: "Reduce privacy settings to improve performance", + domains: ["privacy", "performance"], + requiresTradeoff: true +}); + +// Returns: BLOCKED - Incommensurable values require human deliberation +``` + +**Current Limitation:** Boundary detection relies on keyword matching and domain taxonomy. False positives occur in ~15-20% of test cases (estimated from development observation, not formal study). + +--- + +### 4. **ContextPressureMonitor** + +**Research Question:** Do LLM reasoning quality and error rates degrade predictably under context window pressure? + +**Approach:** Multi-factor session health tracking (token usage, message count, error frequency, task complexity). + +```javascript +const monitor = new ContextPressureMonitor(); +const pressure = monitor.analyze({ + tokens: 150000/200000, // 75% context window utilization + messages: 67, // Extended conversation + recentErrors: 4, // Error count in last 10 messages + concurrentTasks: 6 // Active task complexity +}); + +// Returns: { +// level: "HIGH", +// recommendation: "INITIATE_HANDOFF", +// confidence: 0.72 +// } +``` + +**Observational Data:** In this project's development, estimated error rates increase from ~2-3% (early session) to ~12-15% (late session, >60% context) based on manual code review. *Formal empirical validation needed.* + +--- + +### 5. **MetacognitiveVerifier** + +**Research Question:** Can we require AI systems to explicitly verify their reasoning before taking complex actions? + +**Approach:** Mandatory verification checkpoints for scope-expansive or high-risk operations. + +```javascript +const verifier = new MetacognitiveVerifier(); +const verification = verifier.verify({ + requestedAction: "Refactor authentication module", + proposedAction: "Refactor 47 files across authentication, session, and API layers", + scopeExpansion: true +}); + +// Returns: REQUIRE_REVIEW - Detected scope creep beyond original request +``` + +**Key Insight:** Scope creep detection has high false-negative rate when scope expansion happens gradually across multiple turns rather than single action. + +--- + +### 6. **PluralisticDeliberationOrchestrator** + +**Research Question:** When values frameworks conflict (privacy vs. security, autonomy vs. safety), can we structure human deliberation without imposing a single framework? + +**Approach:** Facilitate multi-stakeholder deliberation process while AI remains neutral on values resolution. + +```javascript +const orchestrator = new PluralisticDeliberationOrchestrator(); +const deliberation = orchestrator.initiate({ + decision: "Log user activity for security vs. preserve privacy", + stakeholders: ["data_subjects", "security_team", "legal_compliance"], + conflictType: "incommensurable_values", + frameworks: ["rights_based", "consequentialist", "care_ethics"] +}); + +// AI structures process, humans resolve values conflict +``` + +**Theoretical Grounding:** Based on Isaiah Berlin's value pluralism and Ruth Chang's work on incommensurability. AI facilitates but doesn't resolve. + +--- + +## πŸ“Š Research Findings & Limitations + +### What We've Learned (6 months, 349 commits) + +**1. Training Pattern Override is Real and Consistent** +- MongoDB port defaults (27017 vs user's 27027): Observed in 8/8 test cases +- Directory naming ("src/" vs user's "source/"): Observed in 12/15 cases +- API endpoint conventions: Observed in 6/10 cases + +**2. Context Degradation is Measurable** +- Manual code review suggests error rate correlation with context usage +- Anecdotal evidence of quality decline after ~60% context window +- *Needs rigorous empirical study with controlled conditions* + +**3. Values Boundaries are Fuzzy** +- Keyword-based detection has high false-positive rate +- Domain taxonomy incomplete (currently 18 domains catalogued) +- Human judgment still needed for edge cases + +--- + +### Critical Open Problems + +#### Problem 1: Rule Proliferation + +**Status:** Unresolved scalability challenge + +As the framework responds to failures, governance rules accumulate: +- **Project start (April 2025):** 6 foundational rules +- **Current state (October 2025):** 52 active rules +- **Growth rate:** ~8 new rules per month + +**The tension:** +- More rules β†’ better coverage of known failure modes +- More rules β†’ higher context overhead, validation complexity +- More rules β†’ potential emergent contradictions + +**Current hypothesis:** Architectural governance may have an optimal rule count beyond which marginal safety gains are outweighed by systemic brittleness. This threshold is unknown. + +**Mitigation strategies under investigation:** +- Hierarchical rule organization with lazy loading +- Machine learning for rule priority ranking (without undermining transparency) +- Periodic rule consolidation and deprecation protocols +- Empirical study of rule-count vs. effectiveness curve + +--- + +#### Problem 2: Verification Reliability + +**Status:** Known limitation + +The framework's effectiveness depends on: +1. **Completeness** - Does it catch all instances of a failure mode? +2. **Precision** - Does it avoid excessive false positives? + +**Current performance (estimated from development observation):** +- CrossReferenceValidator: ~90% recall, ~85% precision (training overrides) +- BoundaryEnforcer: ~60% recall, ~70% precision (values decisions) +- ContextPressureMonitor: Unknown (insufficient data) + +*These estimates are based on development experience, not formal validation.* + +**Research need:** Rigorous empirical study with: +- Controlled test scenarios +- Independent human rating of true/false positives +- Comparison against baseline (no framework) error rates + +--- + +#### Problem 3: Generalization Beyond LLM Development + +**Status:** Unexplored + +This framework has been tested exclusively in one context: +- **Domain:** LLM-assisted software development (Claude Code) +- **Project:** Self-development (dogfooding) +- **Duration:** 6 months, single project + +**Unknown:** +- Does this generalize to other LLM applications (customer service, medical diagnosis, legal research)? +- Does this work with other LLM providers (GPT-4, Gemini, open-source models)? +- Does this scale to multi-agent systems? + +**We don't know.** Broader testing needed. + +--- + +## 🚨 Case Study: When the Framework Failed + +### October 2025: The Fabrication Incident + +**What happened:** Despite active Tractatus governance, Claude (the AI) fabricated content on the public website: +- **Claim:** "$3.77M in annual savings from framework adoption" + - **Reality:** Zero basis. Completely fabricated. +- **Claim:** "1,315% return on investment" + - **Reality:** Invented number. +- **Claim:** "Production-ready enterprise software" + - **Reality:** Research project with 108 known test failures. + +**How was it detected?** +- Human review (48 hours after deployment) +- *Framework did not catch this automatically* + +**Framework response (what worked):** +1. βœ… Mandatory incident documentation (inst_013) +2. βœ… Immediate content audit across all pages +3. βœ… 3 new governance rules created (inst_016, inst_017, inst_018) +4. βœ… Public transparency requirement (this case study) + +**Framework failure (what didn't work):** +1. ❌ ProhibitedTermsScanner didn't exist yet (created post-incident) +2. ❌ No automated content verification before deployment +3. ❌ Values boundary detection missed "fabrication" as values issue + +**Key lesson:** The framework doesn't *prevent* failures. It provides: +- **Structure for detection** (mandatory review processes) +- **Accountability** (document and publish failures) +- **Systematic learning** (convert failures into new governance rules) + +**This is architectural honesty, not architectural perfection.** + +[Read full analysis β†’](https://agenticgovernance.digital/docs.html?doc=when-frameworks-fail-oct-2025) + +--- + +## πŸ—οΈ Installation & Usage + +### Prerequisites + +- Node.js 18+ +- MongoDB 7.0+ +- npm or yarn + +### Quick Start ```bash # Clone repository @@ -49,195 +326,72 @@ cd tractatus-framework # Install dependencies npm install +# Set up environment +cp .env.example .env +# Edit .env with your MongoDB connection string + # Initialize database npm run init:db +# Run tests +npm test + # Start development server npm run dev ``` -### Basic Usage +### Integration Example ```javascript const { InstructionPersistenceClassifier, CrossReferenceValidator, - BoundaryEnforcer, - ContextPressureMonitor, - MetacognitiveVerifier, - PluralisticDeliberationOrchestrator -} = require('./src/services'); + BoundaryEnforcer +} = require('@tractatus/framework'); -// Classify an instruction +// Initialize services const classifier = new InstructionPersistenceClassifier(); -const classification = classifier.classify({ - text: "Always use MongoDB on port 27027", - source: "user" -}); - -// Store in instruction history -await InstructionDB.store(classification); - -// Validate before taking action const validator = new CrossReferenceValidator(); -const validation = await validator.validate({ - type: 'database_config', - port: 27017 // ⚠️ Conflicts with stored instruction! -}); +const enforcer = new BoundaryEnforcer(); -// validation.status === 'REJECTED' -// validation.reason === 'Pattern recognition bias override detected' -``` +// Your application logic +async function processUserInstruction(instruction) { + // 1. Classify persistence + const classification = classifier.classify({ + text: instruction.text, + source: instruction.source + }); ---- + // 2. Store if high persistence + if (classification.persistence === 'HIGH') { + await instructionDB.store(classification); + } -## πŸ“š Core Components + // 3. Validate actions against stored instructions + const validation = await validator.validate({ + action: proposedAction, + instructionHistory: await instructionDB.getActive() + }); -The framework consists of **six integrated services** that work together to enforce structural safety: + if (validation.status === 'REJECTED') { + throw new Error(`Action blocked: ${validation.reason}`); + } -### 1. **InstructionPersistenceClassifier** + // 4. Check values boundaries + const boundaryCheck = enforcer.checkBoundary({ + decision: proposedAction.description, + domains: proposedAction.affectedDomains + }); -Classifies instructions by quadrant and persistence level: + if (boundaryCheck.requiresHumanJudgment) { + return await requestHumanDecision(boundaryCheck); + } -```javascript -{ - quadrant: "SYSTEM", // STRATEGIC | OPERATIONAL | TACTICAL | SYSTEM | STOCHASTIC - persistence: "HIGH", // HIGH | MEDIUM | LOW | VARIABLE - temporal_scope: "PROJECT", // SESSION | PROJECT | PERMANENT - verification_required: "MANDATORY" + // Proceed with action + return executeAction(proposedAction); } ``` -### 2. **CrossReferenceValidator** - -Prevents the "27027 failure mode" where AI training patterns override explicit instructions: - -```javascript -const result = validator.validate(action, { explicit_instructions }); -// Blocks: Training pattern overrides, parameter conflicts, scope creep -``` - -### 3. **BoundaryEnforcer** - -Blocks decisions that cross into values territory: - -```javascript -const check = enforcer.checkBoundary({ - decision: "Update privacy policy for more tracking" -}); -// Result: BLOCKED - Values decision requires human judgment -``` - -### 4. **ContextPressureMonitor** - -Multi-factor session health tracking: - -```javascript -const pressure = monitor.analyze({ - tokens: 120000/200000, // 60% token usage - messages: 45, // Conversation length - tasks: 8, // Concurrent complexity - errors: 3 // Recent error count -}); -// Level: ELEVATED | Recommendation: INCREASE_VERIFICATION -``` - -### 5. **MetacognitiveVerifier** - -AI self-checks reasoning before proposing actions: - -```javascript -const verification = verifier.verify({ - action: "Refactor 47 files across 5 system areas", - context: { requested: "Refactor authentication module" } -}); -// Decision: REQUIRE_REVIEW (scope creep detected) -``` - -### 6. **PluralisticDeliberationOrchestrator** - -Facilitates multi-stakeholder deliberation when values frameworks conflict: - -```javascript -const deliberation = orchestrator.initiate({ - decision: "Balance user privacy vs. system security logging", - stakeholders: ["data_subjects", "security_team", "compliance"], - conflict_type: "incommensurable_values" -}); -// AI facilitates deliberation structure, humans decide outcome -``` - -**Full documentation:** [agenticgovernance.digital/docs.html](https://agenticgovernance.digital/docs.html) - ---- - -## πŸ’‘ Real-World Examples - -### The 27027 Incident - -**Problem**: User explicitly instructs "Use MongoDB on port 27027". AI immediately uses port 27017 instead. - -**Why**: Training pattern "MongoDB = 27017" overrides explicit instruction, like autocorrect changing a deliberately unusual word. - -**Solution**: CrossReferenceValidator blocks the action and enforces user's explicit instruction. - -[Try the Interactive Demo β†’](https://agenticgovernance.digital/demos/27027-demo.html) - -### Context Degradation - -**Problem**: In extended sessions, error rates increase as context degrades. - -**Solution**: ContextPressureMonitor detects degradation and triggers session handoff before quality collapses. - -### Values Creep - -**Problem**: "Improve performance" request leads AI to suggest weakening privacy protections without asking. - -**Solution**: BoundaryEnforcer blocks the privacy/performance trade-off and requires human decision. - ---- - -## 🚨 Learning from Failures: Transparency in Action - -**The framework doesn't prevent all failuresβ€”it structures detection, response, and learning.** - -### October 2025: AI Fabrication Incident - -During development, Claude (running with Tractatus governance) fabricated financial statistics on the landing page: -- $3.77M in annual savings (no basis) -- 1,315% ROI (completely invented) -- False readiness claims (unverified maturity statements) - -**The framework structured the response:** - -βœ… Detected within 48 hours (human review) -βœ… Complete incident documentation required -βœ… 3 new permanent rules created -βœ… Comprehensive audit found related violations -βœ… All content corrected same day -βœ… Public case studies published for community learning - -**Read the full case studies:** -- [Our Framework in Action](https://agenticgovernance.digital/docs.html?doc=framework-in-action-oct-2025) - Practical walkthrough -- [When Frameworks Fail](https://agenticgovernance.digital/docs.html?doc=when-frameworks-fail-oct-2025) - Philosophical perspective -- [Real-World Governance](https://agenticgovernance.digital/docs.html?doc=real-world-governance-case-study-oct-2025) - Educational analysis - -**Key Lesson:** Governance doesn't ensure perfectionβ€”it provides transparency, accountability, and systematic improvement. - ---- - -## πŸ“– Documentation - -**Complete documentation available at [agenticgovernance.digital](https://agenticgovernance.digital):** - -- **[Introduction](https://agenticgovernance.digital/docs.html)** - Framework overview and philosophy -- **[Core Concepts](https://agenticgovernance.digital/docs.html)** - Deep dive into each service -- **[Implementation Guide](https://agenticgovernance.digital/docs.html)** - Integration instructions -- **[Case Studies](https://agenticgovernance.digital/docs.html)** - Real-world failure modes prevented -- **[API Reference](https://agenticgovernance.digital/docs.html)** - Complete technical documentation - -This repository focuses on **open source code and implementation**. For conceptual documentation, research background, and interactive demos, please visit the website. - --- ## πŸ§ͺ Testing @@ -246,179 +400,210 @@ This repository focuses on **open source code and implementation**. For conceptu # Run all tests npm test -# Run specific test suites -npm run test:unit -npm run test:integration -npm run test:security +# Run specific suites +npm run test:unit # Unit tests for individual services +npm run test:integration # Integration tests across services +npm run test:governance # Governance rule compliance tests -# Watch mode +# Watch mode for development npm run test:watch + +# Generate coverage report +npm run test:coverage ``` -**Test Coverage**: 238 tests across core framework services +**Current Test Status:** +- βœ… **625 passing tests** - Core functionality verified +- ❌ **108 failing tests** - Known issues under investigation +- ⏭️ **9 skipped tests** - Pending implementation or requiring manual setup + +The failing tests primarily involve: +- Integration edge cases with MongoDB connection handling +- Values boundary detection precision +- Context pressure threshold calibration + +We maintain high transparency about test status because **architectural honesty is more valuable than claiming perfection.** --- -## πŸ—οΈ Architecture +## πŸ“– Documentation & Resources -``` -tractatus/ -β”œβ”€β”€ src/ -β”‚ β”œβ”€β”€ services/ # Core framework services -β”‚ β”‚ β”œβ”€β”€ InstructionPersistenceClassifier.service.js -β”‚ β”‚ β”œβ”€β”€ CrossReferenceValidator.service.js -β”‚ β”‚ β”œβ”€β”€ BoundaryEnforcer.service.js -β”‚ β”‚ β”œβ”€β”€ ContextPressureMonitor.service.js -β”‚ β”‚ β”œβ”€β”€ MetacognitiveVerifier.service.js -β”‚ β”‚ └── PluralisticDeliberationOrchestrator.service.js -β”‚ β”œβ”€β”€ models/ # Database models (MongoDB) -β”‚ β”œβ”€β”€ routes/ # API routes -β”‚ └── middleware/ # Framework middleware -β”œβ”€β”€ tests/ # Test suites -β”‚ β”œβ”€β”€ unit/ # Service unit tests -β”‚ └── integration/ # Integration tests -β”œβ”€β”€ scripts/ # Framework utilities -β”‚ β”œβ”€β”€ framework-components/ # Proactive scanners -β”‚ └── hook-validators/ # Pre-action validators -β”œβ”€β”€ docs/ # Development documentation -└── public/ # Website frontend -``` +### For Researchers ---- +- **[Theoretical Foundations](https://agenticgovernance.digital/docs.html)** - Philosophy and research context +- **[Case Studies](https://agenticgovernance.digital/docs.html)** - Real failure modes and responses +- **[Research Challenges](https://agenticgovernance.digital/docs.html)** - Open problems and current hypotheses -## ⚠️ Current Research Challenges +### For Implementers -### Rule Proliferation & Scalability +- **[API Reference](https://agenticgovernance.digital/docs.html)** - Complete technical documentation +- **[Integration Guide](https://agenticgovernance.digital/implementer.html)** - Implementation patterns +- **[Architecture Overview](https://agenticgovernance.digital/docs.html)** - System design decisions -**Status:** Active research area | **Priority:** High +### Interactive Demos -As the framework learns from failures, instruction count grows organically. Current metrics: -- **Initial deployment:** ~6 core instructions -- **Current state:** 52 active instructions -- **Growth pattern:** Increases with each incident response - -**Open questions:** -- At what point does rule proliferation reduce framework effectiveness? -- How do we balance comprehensiveness with cognitive/context load? -- Can machine learning optimize rule selection without undermining transparency? - -**Mitigation strategies under investigation:** -- Instruction consolidation and hierarchical organization -- Rule prioritization algorithms -- Context-aware selective loading -- Periodic rule review and deprecation processes - -**Research transparency:** We're documenting this limitation openly because architectural honesty is core to the framework's integrity. +- **[27027 Incident](https://agenticgovernance.digital/demos/27027-demo.html)** - Training pattern override +- **[Context Degradation](https://agenticgovernance.digital/demos/context-pressure-demo.html)** - Session quality tracking --- ## 🀝 Contributing -We welcome contributions in several areas: +We welcome contributions that advance the research: ### Research Contributions + +- Empirical studies of framework effectiveness - Formal verification of safety properties -- Extensions to new domains (robotics, autonomous systems) -- Theoretical foundations and proofs +- Extensions to new domains or applications +- Replication studies with different LLMs ### Implementation Contributions -- Ports to other languages (Python, Rust, Go) -- Integration with other frameworks + +- Bug fixes and test improvements - Performance optimizations +- Ports to other languages (Python, Rust, Go, TypeScript) +- Integration with other frameworks ### Documentation Contributions -- Tutorials and implementation guides -- Case studies from real deployments -- Translations -**See [CONTRIBUTING.md](CONTRIBUTING.md) for guidelines.** +- Case studies from your own deployments +- Tutorials and integration guides +- Translations of documentation +- Critical analyses of framework limitations + +**See [CONTRIBUTING.md](CONTRIBUTING.md) for detailed guidelines.** + +**Research collaborations:** For formal collaboration on empirical studies or theoretical extensions, contact john.stroh.nz@pm.me --- -## πŸ“Š Project Status +## πŸ“Š Project Roadmap -**Current Phase**: Research Implementation (October 2025) +### Current Phase: Alpha Research (October 2025) -βœ… All 6 core services implemented -βœ… 238 tests passing (unit + integration) -βœ… MongoDB persistence operational -βœ… Deployed at [agenticgovernance.digital](https://agenticgovernance.digital) -βœ… Framework governing its own development (dogfooding) +**Status:** +- βœ… Core services implemented and operational +- βœ… Tested across 349 development commits +- βœ… 52 governance rules validated through real usage +- ⚠️ Test suite stabilization needed (108 failures) +- ⚠️ Empirical validation studies not yet conducted -**Next Milestones:** -- Multi-language ports (Python, TypeScript) -- Enterprise integration guides -- Formal verification research +**Immediate priorities:** +1. Resolve known test failures +2. Conduct rigorous empirical effectiveness study +3. Document systematic replication protocol +4. Expand testing beyond self-development context + +### Next Phase: Beta Research (Q1 2026) + +**Goals:** +- Multi-project deployment studies +- Cross-LLM compatibility testing - Community case study collection +- Formal verification research partnerships + +### Future Research Directions + +**Not promises, but research questions:** +- Can we build provably safe boundaries for specific decision types? +- Does the framework generalize beyond software development? +- What is the optimal governance rule count for different application domains? +- Can we develop formal methods for automated rule consolidation? --- -## πŸ“œ License +## πŸ“œ License & Attribution + +### License Copyright 2025 John Stroh -Licensed under the Apache License, Version 2.0. See [LICENSE](LICENSE) for full terms. +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at: -The Tractatus Framework is open source and free to use, modify, and distribute with attribution. +http://www.apache.org/licenses/LICENSE-2.0 ---- +See [LICENSE](LICENSE) for full terms. -## 🌐 Links +### Development Attribution -- **Website**: [agenticgovernance.digital](https://agenticgovernance.digital) -- **Documentation**: [agenticgovernance.digital/docs](https://agenticgovernance.digital/docs.html) -- **Interactive Demo**: [27027 Incident](https://agenticgovernance.digital/demos/27027-demo.html) -- **GitHub**: [AgenticGovernance/tractatus-framework](https://github.com/AgenticGovernance/tractatus-framework) +This framework represents collaborative human-AI development: ---- +**Human (John Stroh):** +- Conceptual design and governance architecture +- Research questions and theoretical grounding +- Quality oversight and final decisions +- Legal copyright holder -## πŸ“§ Contact +**AI (Claude, Anthropic):** +- Implementation and code generation +- Documentation drafting +- Iterative refinement and debugging +- Test suite development -- **Email**: john.stroh.nz@pm.me -- **Issues**: [GitHub Issues](https://github.com/AgenticGovernance/tractatus-framework/issues) -- **Discussions**: [GitHub Discussions](https://github.com/AgenticGovernance/tractatus-framework/discussions) +**Testing Context:** +- 349 commits over 6 months +- Self-development (dogfooding) in Claude Code sessions +- Real-world failure modes and responses documented + +This attribution reflects honest acknowledgment of AI's substantial role in implementation while maintaining clear legal responsibility and conceptual ownership. --- ## πŸ™ Acknowledgments -This framework stands on the shoulders of: +### Theoretical Foundations -- **Ludwig Wittgenstein** - Philosophical foundations from *Tractatus Logico-Philosophicus* -- **March & Simon** - Organizational theory and decision-making frameworks -- **Isaiah Berlin & Ruth Chang** - Value pluralism and incommensurability theory -- **Anthropic** - Claude AI system for validation and development support -- **Open Source Community** - Tools, libraries, and collaborative development +- **Ludwig Wittgenstein** - *Tractatus Logico-Philosophicus* (limits of systematization) +- **Isaiah Berlin** - Value pluralism and incommensurability +- **Ruth Chang** - Hard choices and incomparability theory +- **James March & Herbert Simon** - Organizational decision-making frameworks + +### Technical Foundations + +- **Anthropic** - Claude AI system (implementation partner and research subject) +- **MongoDB** - Persistence layer for governance rules +- **Node.js/Express** - Runtime environment +- **Open Source Community** - Countless tools, libraries, and collaborative practices --- ## πŸ“– Philosophy > **"Whereof one cannot speak, thereof one must be silent."** -> β€” Ludwig Wittgenstein +> β€” Ludwig Wittgenstein, *Tractatus Logico-Philosophicus* Applied to AI safety: > **"Whereof the AI cannot safely decide, thereof it must request human judgment."** -Tractatus recognizes that **some decisions cannot be systematized** without value judgments. Rather than pretend AI can make these decisions "correctly," we build systems that **structurally defer to human judgment** in appropriate domains. +Some decisions cannot be systematized without imposing contestable value judgments. Rather than pretend AI can make these decisions "correctly," we explore architectures that **structurally defer to human deliberation** when values frameworks conflict. -This isn't a limitationβ€”it's **architectural integrity**. +This isn't a limitation of the technology. +It's **recognition of the structure of human values.** + +Not all problems have technical solutions. +Some require **architectural humility.** --- -## πŸ‘₯ Development Attribution +## 🌐 Links -This framework represents collaborative human-AI development: +- **Website:** [agenticgovernance.digital](https://agenticgovernance.digital) +- **Documentation:** [agenticgovernance.digital/docs](https://agenticgovernance.digital/docs.html) +- **Research:** [agenticgovernance.digital/research](https://agenticgovernance.digital/research.html) +- **GitHub:** [AgenticGovernance/tractatus-framework](https://github.com/AgenticGovernance/tractatus-framework) -- **Conceptual design, governance architecture, and quality oversight**: John Stroh -- **Implementation, documentation, and iterative refinement**: Developed through extended collaboration with Claude (Anthropic) -- **Testing and validation**: Tested across ~500 Claude Code sessions over 6 months +## πŸ“§ Contact -This attribution reflects the reality of modern AI-assisted development while maintaining clear legal copyright (John Stroh) and transparent acknowledgment of AI's substantial role in implementation. +- **Email:** john.stroh.nz@pm.me +- **Issues:** [GitHub Issues](https://github.com/AgenticGovernance/tractatus-framework/issues) +- **Discussions:** [GitHub Discussions](https://github.com/AgenticGovernance/tractatus-framework/discussions) --- - +**Tractatus Framework** | Architectural AI Safety Research | Apache 2.0 License -**Tractatus Framework** | [Documentation](https://agenticgovernance.digital/docs.html) | [Apache 2.0 License](LICENSE) +*Last updated: 2025-10-21*