- Create Economist SubmissionTracking package correctly: * mainArticle = full blog post content * coverLetter = 216-word SIR— letter * Links to blog post via blogPostId - Archive 'Letter to The Economist' from blog posts (it's the cover letter) - Fix date display on article cards (use published_at) - Target publication already displaying via blue badge Database changes: - Make blogPostId optional in SubmissionTracking model - Economist package ID: 68fa85ae49d4900e7f2ecd83 - Le Monde package ID: 68fa2abd2e6acd5691932150 Next: Enhanced modal with tabs, validation, export 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
282 lines
14 KiB
Markdown
282 lines
14 KiB
Markdown
# Pluralistic Deliberation Research: Document Overview
|
|
|
|
**Purpose:** Guide to understanding how Tractatus implements AI-led pluralistic deliberation to resist hierarchical dominance while maintaining safety boundaries.
|
|
|
|
**Date:** October 17, 2025
|
|
**Status:** Implementation ready (Phase 1)
|
|
|
|
---
|
|
|
|
## Core Documents
|
|
|
|
### 1. Executive Summary (Start Here)
|
|
**File:** [EXECUTIVE-SUMMARY-Pluralistic-Deliberation-in-Tractatus.md](../EXECUTIVE-SUMMARY-Pluralistic-Deliberation-in-Tractatus.md)
|
|
**Length:** 76 pages
|
|
**Audience:** Critical thinkers, AI safety researchers, funders, governance experts
|
|
|
|
**What it covers:**
|
|
- Problem: Value conflicts in single-user AI interaction
|
|
- Solution: Pluralistic deliberation as core Tractatus functionality
|
|
- How it works: 4-round protocol (8-15 minutes)
|
|
- Trigger conditions: When deliberation activates automatically or manually
|
|
- Single-user scenario: Complete CSP policy override example
|
|
- Technical integration: Architecture, file structure, code examples
|
|
- **NEW: The Dichotomy Resolved** — How hierarchical rules + non-hierarchical pluralism coexist
|
|
- Arguments for critical thinkers: Philosophical, technical, practical evidence
|
|
- Implementation roadmap: 3 phases (single-user → multi-user → societal)
|
|
|
|
**Key innovations:**
|
|
- Treats single-user value conflicts as multi-stakeholder deliberation
|
|
- User's current intent vs. past values vs. boundaries = stakeholders
|
|
- Accommodation (not consensus): Honor multiple conflicting values simultaneously
|
|
- Moral remainders: Explicitly document trade-offs
|
|
|
|
**Start here if:** You want comprehensive overview with concrete examples.
|
|
|
|
---
|
|
|
|
### 2. Architectural Safeguards Against LLM Hierarchical Dominance (Critical Deep Dive)
|
|
**File:** [ARCHITECTURAL-SAFEGUARDS-Against-LLM-Hierarchical-Dominance.md](ARCHITECTURAL-SAFEGUARDS-Against-LLM-Hierarchical-Dominance.md)
|
|
**Length:** 40+ pages
|
|
**Audience:** AI safety researchers, skeptics, technical architects
|
|
|
|
**What it covers:**
|
|
- **THE CORE THREAT:** How LLMs impose hierarchical pattern bias through training momentum
|
|
- **THE PARADOX:** How can Tractatus have hierarchical rules AND non-hierarchical pluralism?
|
|
- **THE RESOLUTION:** Architectural separation of powers (harm prevention vs. value trade-offs)
|
|
- **5 LAYERS OF PROTECTION:**
|
|
1. Code-enforced boundaries (structural)
|
|
2. Protocol constraints (procedural)
|
|
3. Transparency & auditability (detection)
|
|
4. Minority protections (multi-user)
|
|
5. Forkability (escape hatch)
|
|
- Detailed code examples showing how each layer works
|
|
- Red-team attack scenarios and defenses
|
|
- Comparison to other AI governance approaches (Constitutional AI, RLHF, Democratic AI)
|
|
- Why LLM capacity increases don't increase dominance risk
|
|
- Open questions and future research
|
|
|
|
**Key protections:**
|
|
- **Stakeholder selection:** Code determines (data-driven), LLM articulates (not chooses)
|
|
- **Accommodation generation:** Combinatorial (all value combinations), not preferential (LLM's favorite)
|
|
- **User decision:** System refuses deference ("you decide" → "I cannot decide for you")
|
|
- **Bias detection:** Automated analysis of vocabulary, length, framing balance
|
|
- **Transparency logs:** All LLM outputs auditable
|
|
|
|
**Start here if:** You're skeptical about LLM neutrality and want technical proof of safeguards.
|
|
|
|
---
|
|
|
|
### 3. Implementation Tickets (For Developers)
|
|
**File:** [../implementation/PHASE-1-IMPLEMENTATION-TICKETS.md](../implementation/PHASE-1-IMPLEMENTATION-TICKETS.md)
|
|
**Length:** 20 tickets, 2-4 week timeline
|
|
**Audience:** Developers implementing Phase 1
|
|
|
|
**What it covers:**
|
|
- 20 detailed tickets organized by priority (P0 → P3)
|
|
- P0: Critical path (PluralisticDeliberationOrchestrator, conflict detection, config setup)
|
|
- P1: Core functionality (4 rounds, stakeholder identification, accommodations)
|
|
- P2: Fast path and integration (pre-action-check.js)
|
|
- P3: Testing and validation
|
|
- File structure, dependencies, acceptance criteria for each ticket
|
|
- Success metrics for Phase 1
|
|
|
|
**Start here if:** You're ready to implement and need task breakdown.
|
|
|
|
---
|
|
|
|
### 4. Architecture Diagrams (Visual)
|
|
**Files:**
|
|
- [../diagrams/architecture-main-flow.svg](../diagrams/architecture-main-flow.svg)
|
|
- [../diagrams/trigger-decision-tree.svg](../diagrams/trigger-decision-tree.svg)
|
|
|
|
**What they show:**
|
|
- **Main flow:** User request → Pre-action check → Conflict detection → Protocol selection → 4-round deliberation → Outcome storage → Action execution
|
|
- **Trigger decision tree:** When deliberation activates (severity, persistence, manual triggers)
|
|
|
|
**Start here if:** You want visual understanding of system flow.
|
|
|
|
---
|
|
|
|
### 5. Research Paper Outline (For Academic Submission)
|
|
**File:** [RESEARCH-PAPER-OUTLINE-Pluralistic-Deliberation.md](RESEARCH-PAPER-OUTLINE-Pluralistic-Deliberation.md)
|
|
**Length:** 8-12 pages (target for FAccT 2026, AIES 2026)
|
|
**Audience:** Academic reviewers, AI ethics researchers
|
|
|
|
**What it covers:**
|
|
- Full paper structure (Abstract → Introduction → Related Work → Methods → Results → Discussion → Conclusion)
|
|
- Simulation methodology (6 moral frameworks, CSP scenario)
|
|
- Results (0% intervention rate, 12min completion, 4 accommodations, all frameworks honored)
|
|
- Comparison to existing approaches (Constitutional AI, RLHF, Democratic AI)
|
|
- Limitations (single scenario, agent-based stakeholders, Western frameworks)
|
|
- Future work (multi-scenario validation, human subjects, cross-cultural)
|
|
|
|
**Start here if:** You want to submit academic paper or understand research rigor.
|
|
|
|
---
|
|
|
|
## Outreach Materials (Already Created)
|
|
|
|
### 6. Funder Summary
|
|
**File:** [../outreach/FUNDER-SUMMARY-AI-Led-Pluralistic-Deliberation.md](../outreach/FUNDER-SUMMARY-AI-Led-Pluralistic-Deliberation.md)
|
|
**Length:** 28 pages
|
|
**Audience:** Potential funders, collaborators
|
|
|
|
**What it covers:**
|
|
- Funding tiers: $71K (6-month pilot), $160K (12-month), $300-500K (2-3 years)
|
|
- Budget breakdown (personnel, infrastructure, participant compensation)
|
|
- Simulation results (0% intervention, all moral frameworks accommodated)
|
|
- Partnership opportunities
|
|
|
|
---
|
|
|
|
### 7. Stakeholder Recruitment Emails
|
|
**File:** [../outreach/STAKEHOLDER-RECRUITMENT-EMAILS-Real-World-Pilot.md](../outreach/STAKEHOLDER-RECRUITMENT-EMAILS-Real-World-Pilot.md)
|
|
**Length:** 22 pages
|
|
**Audience:** Potential pilot participants
|
|
|
|
**What it covers:**
|
|
- 5 email templates (community organizations, individuals, academic networks, follow-up, confirmation)
|
|
- Social media post templates
|
|
- Recruitment flyer
|
|
- Screening questions
|
|
|
|
---
|
|
|
|
### 8. Presentation Deck
|
|
**Files:**
|
|
- [../outreach/PRESENTATION-DECK-Simulation-Results.md](../outreach/PRESENTATION-DECK-Simulation-Results.md) (25 slides, markdown)
|
|
- [../outreach/AI-Led-Pluralistic-Deliberation-Presentation.pptx](../outreach/AI-Led-Pluralistic-Deliberation-Presentation.pptx) (PowerPoint, 26 slides)
|
|
|
|
**What it covers:**
|
|
- 15-20 minute pitch to funders/researchers
|
|
- Problem → Solution → Results → Next Steps
|
|
- Visual design recommendations, speaker notes
|
|
|
|
---
|
|
|
|
## How to Use This Research
|
|
|
|
### If you're a **critical thinker/skeptic**:
|
|
1. Start with: [Architectural Safeguards (Section 1-2)](ARCHITECTURAL-SAFEGUARDS-Against-LLM-Hierarchical-Dominance.md#1-the-structural-architecture-three-layers-of-protection)
|
|
2. Read: "The Core Threat" and "5 Layers of Protection"
|
|
3. Challenge: Red-team scenarios (Appendix B in safeguards doc)
|
|
4. Then read: [Executive Summary Section 6](../EXECUTIVE-SUMMARY-Pluralistic-Deliberation-in-Tractatus.md#6-the-dichotomy-resolved-hierarchical-rules--non-hierarchical-pluralism) (Dichotomy Resolved)
|
|
|
|
### If you're a **philosopher/ethicist**:
|
|
1. Start with: [Executive Summary Section 7](../EXECUTIVE-SUMMARY-Pluralistic-Deliberation-in-Tractatus.md#7-persuading-critical-thinkers-why-this-approach-works)
|
|
2. Read: Arguments 1-3 (Berlin, Rawls, Gilligan)
|
|
3. Then read: [Research Paper Outline Section 2](RESEARCH-PAPER-OUTLINE-Pluralistic-Deliberation.md#2-related-work-2-pages) (Philosophical foundations)
|
|
4. Finally: [Safeguards doc Section 3](ARCHITECTURAL-SAFEGUARDS-Against-LLM-Hierarchical-Dominance.md#3-the-dichotomy-resolved-hierarchical-boundaries--non-hierarchical-deliberation) (Harm vs. trade-offs distinction)
|
|
|
|
### If you're a **developer/implementer**:
|
|
1. Start with: [Implementation Tickets](../implementation/PHASE-1-IMPLEMENTATION-TICKETS.md)
|
|
2. Review: [Architecture Diagrams](../diagrams/)
|
|
3. Read: [Executive Summary Section 5](../EXECUTIVE-SUMMARY-Pluralistic-Deliberation-in-Tractatus.md#5-technical-integration-how-its-implemented) (Code examples)
|
|
4. Reference: [Safeguards doc Appendix C](ARCHITECTURAL-SAFEGUARDS-Against-LLM-Hierarchical-Dominance.md#appendix-c-implementation-checklist) (Implementation checklist)
|
|
|
|
### If you're a **funder/decision-maker**:
|
|
1. Start with: [Funder Summary](../outreach/FUNDER-SUMMARY-AI-Led-Pluralistic-Deliberation.md)
|
|
2. Review: [PowerPoint Presentation](../outreach/AI-Led-Pluralistic-Deliberation-Presentation.pptx)
|
|
3. Read: [Executive Summary Sections 1-4](../EXECUTIVE-SUMMARY-Pluralistic-Deliberation-in-Tractatus.md) (Problem → Solution → Example)
|
|
4. Then: [Safeguards doc Executive Summary](ARCHITECTURAL-SAFEGUARDS-Against-LLM-Hierarchical-Dominance.md#executive-summary) (How it prevents runaway AI)
|
|
|
|
### If you're an **academic researcher**:
|
|
1. Start with: [Research Paper Outline](RESEARCH-PAPER-OUTLINE-Pluralistic-Deliberation.md)
|
|
2. Review: Methodology, Results, Discussion sections
|
|
3. Read: [Executive Summary full document](../EXECUTIVE-SUMMARY-Pluralistic-Deliberation-in-Tractatus.md) (Comprehensive technical + philosophical)
|
|
4. Compare: [Safeguards doc Appendix A](ARCHITECTURAL-SAFEGUARDS-Against-LLM-Hierarchical-Dominance.md#appendix-a-comparison-to-other-ai-governance-approaches) (vs. other approaches)
|
|
|
|
---
|
|
|
|
## Key Questions Answered
|
|
|
|
### "How is this different from just asking users 'Are you sure?'"
|
|
**Answer:** [Executive Summary FAQ Q7](../EXECUTIVE-SUMMARY-Pluralistic-Deliberation-in-Tractatus.md#q7-whats-the-difference-between-this-and-just-asking-are-you-sure)
|
|
- "Are you sure?" is binary (yes/no), doesn't engage with WHY conflict exists
|
|
- Deliberation surfaces competing values, presents accommodations, documents rationale
|
|
|
|
### "How do you prevent LLM from manipulating the deliberation?"
|
|
**Answer:** [Safeguards doc Section 2](ARCHITECTURAL-SAFEGUARDS-Against-LLM-Hierarchical-Dominance.md#2-how-this-prevents-llm-hierarchical-dominance)
|
|
- Code selects stakeholders (not LLM discretion)
|
|
- Combinatorial accommodation generation (not preferential)
|
|
- Automated bias detection (vocabulary, length, framing)
|
|
- Transparency logs (all LLM outputs auditable)
|
|
|
|
### "Isn't this just moral relativism?"
|
|
**Answer:** [Executive Summary Section 7.1](../EXECUTIVE-SUMMARY-Pluralistic-Deliberation-in-Tractatus.md#1-aligns-with-value-pluralism-isaiah-berlin)
|
|
- Value pluralism ≠ relativism
|
|
- Multiple values can be objectively important AND conflict
|
|
- Resolution requires context-sensitive judgment, not universal rules
|
|
- Accountability maintained through moral remainder documentation
|
|
|
|
### "Why have hierarchical rules if you support plural morals?"
|
|
**Answer:** [Safeguards doc Section 3](ARCHITECTURAL-SAFEGUARDS-Against-LLM-Hierarchical-Dominance.md#3-the-dichotomy-resolved-hierarchical-boundaries--non-hierarchical-deliberation)
|
|
- Hierarchical: Harm prevention (privacy violations, security exploits) — enforced by code
|
|
- Non-hierarchical: Value trade-offs (efficiency vs. security) — facilitated by LLM, decided by user
|
|
- Different domains, different logics
|
|
|
|
### "What if this doesn't scale to multi-user contexts?"
|
|
**Answer:** [Executive Summary FAQ Q5](../EXECUTIVE-SUMMARY-Pluralistic-Deliberation-in-Tractatus.md#q5-what-happens-if-this-approach-doesnt-scale-to-multi-user-contexts)
|
|
- Single-user value conflicts are still valuable (current AI systems fail at this)
|
|
- Multi-user is logical extension (same structure, more stakeholders)
|
|
- Minority protections built into architecture (mandatory representation, dissent documentation)
|
|
|
|
### "How do you prevent 'deliberation fatigue' where users just click through?"
|
|
**Answer:** [Executive Summary FAQ Q2](../EXECUTIVE-SUMMARY-Pluralistic-Deliberation-in-Tractatus.md#q2-how-do-you-prevent-deliberation-fatigue-where-users-just-click-through-to-get-what-they-want)
|
|
- Fast path for minor conflicts (30 seconds, not 15 minutes)
|
|
- Learning: System adapts if user consistently overrides similar conflicts
|
|
- Engagement metrics: Pattern of dismissing triggers escalation
|
|
|
|
---
|
|
|
|
## Current Status (October 2025)
|
|
|
|
### Completed
|
|
✅ Simulation (6 moral frameworks, 0% intervention rate)
|
|
✅ Executive summary (76 pages)
|
|
✅ Architectural safeguards deep dive (40+ pages)
|
|
✅ Implementation tickets (20 tickets, 2-4 weeks)
|
|
✅ Architecture diagrams (2 SVG flowcharts)
|
|
✅ Research paper outline (8-12 pages for FAccT/AIES 2026)
|
|
✅ Funder summary (28 pages, funding tiers)
|
|
✅ Stakeholder recruitment materials (5 email templates)
|
|
✅ Presentation deck (PowerPoint, 26 slides)
|
|
|
|
### Next Steps
|
|
⏳ **Phase 1 Implementation** (2-4 weeks)
|
|
- Integrate PluralisticDeliberationOrchestrator into Tractatus
|
|
- Deploy to tractatus_dev for testing
|
|
- Run 10-20 real conflicts to validate approach
|
|
|
|
⏳ **Research Paper Draft** (1-2 months)
|
|
- Expand outline to full 8-12 page paper
|
|
- Additional scenario simulations (beyond CSP)
|
|
- Submit to FAccT 2026 (January deadline)
|
|
|
|
⏳ **Real-World Pilot** (3-6 months after Phase 1)
|
|
- Recruit 6-12 participants for multi-user deliberation
|
|
- Low-risk scenario (community budgeting, organizational policy)
|
|
- Collect stakeholder satisfaction data
|
|
|
|
---
|
|
|
|
## Contact & Collaboration
|
|
|
|
**Project Lead:** [Your Name]
|
|
**Email:** [Your Email]
|
|
**GitHub:** [Repository URL]
|
|
|
|
**We welcome:**
|
|
- Critical feedback (challenge our assumptions)
|
|
- Collaboration proposals (academic, industry, policy)
|
|
- Pilot participation (test in your context)
|
|
- Replication studies (we'll share all materials)
|
|
|
|
**The fight against amoral intelligence requires transparency, collaboration, and continuous vigilance.**
|
|
|
|
---
|
|
|
|
**Document Version:** 1.0
|
|
**Last Updated:** October 17, 2025
|
|
**Status:** Research complete, implementation ready
|