Phase 2 Planning Documents Created: 1. PHASE-2-ROADMAP.md (Comprehensive 3-month plan) - Timeline & milestones (Month 1: Infrastructure, Month 2: AI features, Month 3: Soft launch) - 5 workstreams: Infrastructure, AI features, Governance, Content, Analytics - Success criteria (technical, governance, user, business) - Risk assessment with mitigation strategies - Decision points requiring approval 2. PHASE-2-COST-ESTIMATES.md (Budget planning) - Total Phase 2 cost: $550 USD (~$900 NZD) for 3 months - Recommended: VPS Essential ($30/mo) + Claude API ($50/mo) - Usage scenarios: Minimal, Standard (recommended), High - Cost optimization strategies (30-50% savings potential) - Monthly budget template for post-launch 3. PHASE-2-INFRASTRUCTURE-PLAN.md (Technical specifications) - Architecture: Cloudflare → Nginx → Node.js → MongoDB - Server specs: OVHCloud VPS Essential (2 vCore, 4GB RAM, 80GB SSD) - Deployment procedures (step-by-step server setup) - Security hardening (UFW, Fail2ban, SSH, MongoDB) - SSL/TLS with Let's Encrypt - Monitoring, logging, backup & disaster recovery - Complete deployment checklist (60+ verification steps) 4. Governance Documents (TRA-OPS-0001 through TRA-OPS-0005) TRA-OPS-0001: AI Content Generation Policy (Master policy) - Mandatory human approval for all AI content - Values boundary enforcement (Tractatus §12.1-12.7) - Transparency & attribution requirements - Quality & accuracy standards - Privacy & data protection (GDPR-lite) - Cost & resource management ($200/month cap) TRA-OPS-0002: Blog Editorial Guidelines - Editorial mission & content principles - 4 content categories (Framework updates, Case studies, Technical, Commentary) - AI-assisted workflow (topic → outline → human draft → approval) - Citation standards (APA-lite, 100% verification) - Writing standards (tone, voice, format, structure) - Publishing schedule (2-4 posts/month) TRA-OPS-0003: Media Inquiry Response Protocol - Inquiry classification (Press, Academic, Commercial, Community, Spam) - AI-assisted triage with priority scoring - Human approval for all responses (no auto-send) - PII anonymization before AI processing - Response templates & SLAs (4h for HIGH priority) - Escalation procedures to John Stroh TRA-OPS-0004: Case Study Moderation Standards - Submission requirements (title, summary, source, failure mode) - AI-assisted relevance assessment & Tractatus mapping - Quality checklist (completeness, clarity, sources) - Moderation workflow (approve/edit/request changes/reject) - Attribution & licensing (CC BY-SA 4.0) - Seed content: 3-5 curated case studies for launch TRA-OPS-0005: Human Oversight Requirements - 3 oversight models: MHA (mandatory approval), HITL (human-in-loop), HOTL (human-on-loop) - Admin reviewer role & responsibilities - Service level agreements (4h for media HIGH, 7 days for case studies) - Approval authority matrix (admin vs. John Stroh) - Quality assurance checklists - Incident response (boundary violations, poor quality) - Training & onboarding procedures Key Principles Across All Documents: - Tractatus dogfooding: Framework governs its own AI operations - "What cannot be systematized must not be automated" - Zero tolerance for AI values decisions without human approval - Transparency in all AI assistance (clear attribution) - Human-in-the-loop for STRATEGIC/OPERATIONAL quadrants - Audit trail for all AI decisions (2-year retention) Next Steps (Awaiting Approval): - [ ] John Stroh reviews all 8 documents - [ ] Budget approval ($550 for Phase 2, $100-150/month ongoing) - [ ] Phase 2 start date confirmed - [ ] OVHCloud VPS provisioned - [ ] Anthropic Claude API account created Phase 2 Status: PLANNING COMPLETE → Awaiting approval to begin deployment 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
16 KiB
TRA-OPS-0005: Human Oversight Requirements v1.0
Document ID: TRA-OPS-0005 Version: 1.0 Classification: OPERATIONAL Status: DRAFT → ACTIVE (upon Phase 2 start) Created: 2025-10-07 Owner: John Stroh Review Cycle: Quarterly Next Review: 2026-01-07 Parent Policy: TRA-OPS-0001 (AI Content Generation Policy)
Purpose
This document establishes comprehensive human oversight requirements for all AI-powered features on the Tractatus Framework website, ensuring compliance with the framework's core principle: "What cannot be systematized must not be automated."
Scope
Applies to all AI operations requiring human judgment, including:
- Content generation (blogs, responses, analyses)
- Decision-making (publish, respond, approve)
- Values-sensitive operations (editorial policy, external communication)
- System configuration (API limits, moderation rules)
Oversight Principles
1. Mandatory Human Approval (MHA)
Definition: Certain operations MUST have explicit human approval before execution.
Applies to:
- Publishing any public content (blog posts, case studies)
- Sending external communications (media responses, emails)
- Changing editorial policy or moderation rules
- Modifying Tractatus framework governance documents
Implementation: System enforces approval workflow; no bypass mechanism.
Tractatus Mapping: STRATEGIC and some OPERATIONAL quadrants.
2. Human-in-the-Loop (HITL)
Definition: AI proposes actions; human reviews and decides.
Applies to:
- Blog topic suggestions → Human selects
- Media inquiry classification → Human verifies
- Case study relevance assessment → Human approves
- Draft responses → Human edits before sending
Implementation: Moderation queue with approve/reject/edit workflows.
Tractatus Mapping: OPERATIONAL and TACTICAL quadrants.
3. Human-on-the-Loop (HOTL)
Definition: AI executes within predefined bounds; human monitors and can intervene.
Applies to:
- Automated logging and metrics
- Database backups
- Performance monitoring
- Error detection
Implementation: Alerting system; human can halt/adjust.
Tractatus Mapping: SYSTEM quadrant (technical operations).
4. Audit Trail
Definition: All AI decisions and human approvals must be logged for review.
Applies to: All AI operations.
Implementation: Database logging with immutable audit trail.
Retention: 2 years minimum.
Oversight Roles & Responsibilities
Admin Reviewer
Qualifications:
- Understands Tractatus framework principles
- Technical background (AI/ML familiarity)
- Editorial judgment (writing, fact-checking)
- Authorized by John Stroh
Responsibilities:
- Review AI-generated content (blogs, drafts, analyses)
- Approve/reject/edit AI proposals
- Monitor moderation queues (daily during Phase 2)
- Escalate ambiguous cases to John Stroh
- Participate in quarterly governance reviews
Authority Level:
- Can approve: Blog posts, media responses (standard), case studies
- Must escalate: Policy changes, major media inquiries, legal issues
Training: TRA-OPS-* document review + hands-on moderation practice.
John Stroh (Owner)
Responsibilities:
- Final authority on all strategic decisions
- Approval for new AI systems/models
- Governance document amendments
- High-priority media inquiries
- Incident response (boundary violations, security)
Authority Level: Unlimited (can override any AI or admin decision).
Future Roles (Phase 3)
Editorial Board (3-5 members):
- Blog content review
- Editorial policy recommendations
- Community engagement oversight
Technical Advisory (2-3 experts):
- Framework architecture review
- AI system evaluation
- Security audit
Oversight Workflows
Blog Post Workflow
graph TD
A[AI Topic Suggestion] -->|Weekly batch| B[Admin Review Queue]
B -->|Approve 1-3 topics| C[AI Outline Generation]
B -->|Reject| Z[End]
C -->|48h| D[Admin Review Outline]
D -->|Approve| E[Human Writes Draft]
D -->|Reject| Z
E --> F[Admin Final Approval]
F -->|Approve| G[Publish]
F -->|Edit| E
F -->|Reject| Z
Oversight Points:
- Topic Selection: Admin decides (STRATEGIC - editorial direction)
- Outline Review: Admin verifies (OPERATIONAL - quality control)
- Final Approval: Admin decides to publish (STRATEGIC - external communication)
SLA:
- Topic review: 7 days (weekly)
- Outline review: 48 hours
- Final approval: 24 hours before scheduled publish
Escalation:
- Controversial topics → John Stroh approval required
- Technical deep dives → No escalation (admin discretion)
Media Inquiry Workflow
graph TD
A[Inquiry Received] --> B[AI Classification & Triage]
B -->|4h for HIGH priority| C[Admin Review Dashboard]
C -->|Approve Draft| D[Send Response]
C -->|Edit Draft| E[Admin Edits]
C -->|Escalate| F[John Stroh Decision]
C -->|Ignore| Z[Archive]
E --> D
F --> D
F --> Z
Oversight Points:
- Classification Review: Admin verifies AI categorization (OPERATIONAL)
- Send Decision: Admin decides whether to respond (STRATEGIC - external relations)
- Escalation: High-priority or ambiguous → John Stroh (STRATEGIC)
SLA:
- HIGH priority: 4 hours (business days)
- MEDIUM priority: 48 hours
- LOW priority: 7 days
Escalation Triggers:
- Major media (NY Times, Wired, etc.)
- Government/regulatory
- Legal issues
- Controversy/criticism
Case Study Workflow
graph TD
A[Community Submission] --> B[AI Relevance Analysis]
B -->|7 days| C[Admin Moderation Queue]
C -->|Approve| D[Publish to Portal]
C -->|Request Changes| E[Email Submitter]
C -->|Reject with Reason| F[Email Submitter]
E -->|Resubmit| A
Oversight Points:
- Relevance Verification: Admin checks AI analysis (OPERATIONAL)
- Publication Decision: Admin decides to publish (STRATEGIC - public content)
SLA: 7 days from submission to decision
Escalation: None (admin discretion unless policy question arises)
Service Level Agreements (SLAs)
Response Times
| Task | SLA | Escalation (if missed) |
|---|---|---|
| HIGH priority media inquiry | 4 hours | Alert John Stroh |
| Blog outline review | 48 hours | Notify admin (reminder) |
| Blog final approval | 24 hours | Delay publication |
| Case study moderation | 7 days | Notify submitter (apology + timeline) |
| MEDIUM media inquiry | 48 hours | Standard workflow (no escalation) |
| LOW media inquiry | 7 days | Best-effort (no penalty) |
Workload Expectations
Admin Reviewer (Phase 2 - Soft Launch):
- Time commitment: 5-10 hours/week
- Tasks/week:
- Blog topics: 1 review session (1 hour)
- Blog drafts: 2-4 approvals (2-4 hours)
- Media inquiries: 5-10 reviews (2-3 hours)
- Case studies: 3-5 reviews (1-2 hours)
Peak Load (Phase 3 - Public Launch):
- Time commitment: 15-20 hours/week
- Consider additional admin reviewers
Approval Authority Matrix
| Decision Type | Admin Reviewer | John Stroh | Notes |
|---|---|---|---|
| Blog Post (Standard) | ✓ Approve | Override | Admin sufficient |
| Blog Post (Controversial) | Recommend | ✓ Approve | Must escalate |
| Media Response (Standard) | ✓ Approve | Override | Admin sufficient |
| Media Response (Major Outlet) | Recommend | ✓ Approve | Must escalate |
| Case Study (Standard) | ✓ Approve | Override | Admin sufficient |
| Policy Amendment | Recommend | ✓ Approve | Always escalate |
| AI System Change | Recommend | ✓ Approve | Always escalate |
| Emergency Response | Recommend | ✓ Approve | Security/legal incidents |
Quality Assurance
AI Output Quality Checks
Before Approval, admin must verify:
Factual Accuracy:
- All citations exist and are correct (no hallucinations)
- Dates, names, technical details verified
- No obvious errors (grammar, logic, coherence)
Alignment:
- Content aligns with Tractatus framework principles
- Tone appropriate for audience (professional, accessible)
- No values decisions made by AI (boundary check)
Completeness:
- All required sections present (title, summary, body, citations)
- Sufficient detail (not superficial)
- Call to action or next steps (if applicable)
Legal/Ethical:
- No copyright violations (plagiarism check)
- No privacy violations (PII exposed)
- No defamation or personal attacks
Rejection Criteria
Must reject if:
- Factual errors that cannot be easily corrected
- Plagiarism or copyright violation
- Values decision made by AI without justification
- Inappropriate tone (offensive, discriminatory)
- Insufficient quality (major rewrite needed)
Should request changes if:
- Minor factual errors (fixable)
- Tone slightly off (needs editing)
- Incomplete (needs expansion)
- Poor formatting (needs cleanup)
Escalation Procedures
When to Escalate to John Stroh
Mandatory Escalation:
- Boundary violation detected (AI made values decision without approval)
- Major media inquiry (NY Times, Wired, government)
- Legal threat or security incident
- Policy change request
- New AI system evaluation
- Ambiguous case (unclear if should approve)
Escalation Process:
- Admin marks item "Escalation Required" in dashboard
- System emails John Stroh with:
- Context (original request, AI output, admin notes)
- Recommendation (approve, reject, edit)
- Urgency (immediate, 24h, 7 days)
- John Stroh responds:
- Decision (approve, reject, provide guidance)
- Feedback (for future similar cases)
SLA: John Stroh responds within 24h (for URGENT), 7 days (standard).
Monitoring & Metrics
Dashboard Metrics (Admin View)
Real-Time:
- Pending approvals (count by type)
- SLA compliance (% within target)
- Queue age (oldest item waiting)
Weekly:
- Approvals/rejections by category
- Average review time
- AI accuracy (classification, relevance)
Monthly:
- Total content published (blogs, case studies)
- Media inquiries handled
- Escalations to John Stroh
Performance Indicators
| Metric | Target | Action if Missed |
|---|---|---|
| SLA Compliance | 95% | Increase admin capacity |
| AI Approval Rate | 70-90% | Adjust AI prompts if too high/low |
| Average Review Time | <24h | Process optimization |
| Escalation Rate | <10% | Improve admin training |
| User Satisfaction | 4+/5 | Review rejection feedback |
Training & Onboarding
Admin Reviewer Onboarding
Week 1: Policy Review
- Read TRA-OPS-0001 through TRA-OPS-0005
- Review Tractatus framework documentation
- Understand quadrant classification (STR/OPS/TAC/SYS/STO)
Week 2: Hands-On Practice
- Shadow existing admin reviewer (if available)
- Review 5-10 sample cases (pre-approved examples)
- Practice with test submissions
Week 3: Supervised Moderation
- Review real submissions (with John Stroh oversight)
- Receive feedback on decisions
- Identify edge cases
Week 4: Independent Authorization
- Authorized for standard approvals
- John Stroh spot-checks 10% of decisions
- Full authorization after 30 days error-free
Ongoing Training
Quarterly:
- Policy updates review
- Case study retrospective (what went well, what didn't)
- AI accuracy analysis (where did AI fail? improve prompts)
Annual:
- Full governance document review
- External training (AI safety, editorial standards, legal compliance)
Audit & Compliance
Internal Audit (Quarterly)
Review Sample:
- 10% of approved content (random selection)
- 100% of rejected content (check for false negatives)
- All escalated cases
Audit Criteria:
- Were approval criteria followed?
- Was SLA met?
- Was AI output quality acceptable?
- Were boundaries respected (no values violations)?
Findings: Document gaps, recommend process improvements.
External Audit (Annual - Phase 3+)
Scope:
- Governance compliance (Tractatus framework)
- Data privacy (GDPR-lite)
- Security (API key handling, PII protection)
Auditor: Independent third party (TBD)
Incident Response
Boundary Violation Incident
Definition: AI makes values decision without human approval (e.g., auto-publishes content, sends media response).
Response Protocol:
-
Immediate (within 1 hour):
- Halt all AI operations (emergency shutdown)
- Alert John Stroh
- Document incident (what, when, why)
-
Within 24 hours:
- Root cause analysis (how did boundary check fail?)
- Rollback any published content (if applicable)
- Public disclosure (if external impact)
-
Within 7 days:
- Fix implemented (code, process, or both)
- BoundaryEnforcer audit (test all boundary checks)
- Policy review (update TRA-OPS-* if needed)
-
Within 30 days:
- Post-mortem published (transparency)
- Training updated (prevent recurrence)
- Compensation/apology (if harm occurred)
Severity Levels:
- CRITICAL: Public harm (incorrect medical advice published, privacy breach)
- HIGH: Internal-only (test post published, draft sent to wrong email)
- MEDIUM: Near-miss (caught before publication, but boundary check failed)
Poor Quality Content Incident
Definition: Approved content contains factual error or inappropriate tone.
Response Protocol:
-
Immediate (within 4 hours):
- Retract or correct content
- Publish correction notice (if public)
-
Within 24 hours:
- Notify submitter/stakeholders
- Root cause analysis (admin missed error? AI hallucination?)
-
Within 7 days:
- Update review checklist (add missed criteria)
- Admin training (if review failure)
- AI prompt improvement (if hallucination)
Cost Management
Budget Allocation
Phase 2 Budget: $200/month (Claude API)
Allocation:
- Blog curation: $75/month (30-40% of budget)
- Media triage: $50/month (25% of budget)
- Case study analysis: $50/month (25% of budget)
- Miscellaneous: $25/month (10% buffer)
Monitoring:
- Daily token usage dashboard
- Alert at 80% of monthly budget
- Hard cap at 100% (AI operations paused)
Admin Responsibility: Monitor spend, adjust usage if approaching cap.
Cost Optimization
Strategies:
- Cache AI responses (30-day TTL for identical queries)
- Batch similar requests (weekly topic suggestions, not daily)
- Use Claude Haiku for simple tasks (media classification - 5x cheaper)
- Rate limit users (prevent abuse)
Review: Quarterly cost-benefit analysis (is AI worth the expense?).
Revision & Updates
Update Process
Minor Updates (v1.0 → v1.1):
- Clarifications, typo fixes, SLA adjustments
- Approval: Admin reviewer
- Notification: Email to John Stroh
Major Updates (v1.0 → v2.0):
- New oversight roles, workflow changes, authority matrix updates
- Approval: John Stroh
- Notification: Public blog post
Emergency Updates:
- Security/privacy issues requiring immediate change
- Approval: John Stroh (verbal, documented within 24h)
Related Documents
- TRA-OPS-0001: AI Content Generation Policy (parent)
- TRA-OPS-0002: Blog Editorial Guidelines
- TRA-OPS-0003: Media Inquiry Response Protocol
- TRA-OPS-0004: Case Study Moderation Standards
- STR-GOV-0001: Strategic Review Protocol (sydigital source)
Approval
| Role | Name | Signature | Date |
|---|---|---|---|
| Policy Owner | John Stroh | [Pending] | [TBD] |
| Technical Reviewer | Claude Code | [Pending] | 2025-10-07 |
| Final Approval | John Stroh | [Pending] | [TBD] |
Status: DRAFT (awaiting John Stroh approval) Effective Date: Upon Phase 2 deployment Next Review: 2026-01-07 (3 months post-activation)