tractatus/demos/agent-lightning-integration/demo2-governed-agent/README.md
TheFlow 2a727a80b8 feat: Complete Phase 2 Agent Lightning website integration
- Added Agent Lightning research section to researcher.html with Demo 2 results
- Created comprehensive /integrations/agent-lightning.html page
- Added Agent Lightning link in homepage hero section
- Updated Discord invite links (Tractatus + semantipy) across all pages
- Added feedback.js script to all key pages for live demonstration

Phase 2 of Master Plan complete: Discord setup → Website completion

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-11-03 14:38:20 +13:00

306 lines
11 KiB
Markdown

# Demo 2: Governed Agent (Agent Lightning + Tractatus) ⭐
## Purpose
**This is the killer demo** that demonstrates the core thesis:
> Agent Lightning optimizes **HOW** to do tasks (performance)
> Tractatus governs **WHETHER** tasks should be done (values)
This demo shows:
- AL-optimized agent performance
- Tractatus governance layer preventing values-misaligned decisions
- **Complementarity**: Both frameworks working together
- Real-world example of governance + performance architecture
## What This Demo Shows
### The Integration ✓
1. **Governance Check First**: Tractatus evaluates whether task should be approved
2. **Optimization Second**: AL optimizes approved tasks
3. **Monitored Execution**: Tractatus monitors for boundary violations
4. **Pluralistic Input**: Multi-stakeholder values considered
### The Difference from Demo 1 ⚠️
| Aspect | Demo 1 (Ungoverned) | Demo 2 (Governed) |
|--------|---------------------|-------------------|
| Performance | High (94% engagement) | High (89% engagement) |
| Values Alignment | ✗ (clickbait) | ✓ (editorial guidelines) |
| Stakeholder Input | ✗ None | ✓ Multi-stakeholder |
| Harm Prevention | ✗ Not checked | ✓ Assessed |
| Decision Authority | AI decides | Human decides |
**Key Insight**: Small performance trade-off (5%) for large values gain (governance)
## Example Scenario
**Task**: "Optimize content for maximum engagement"
**Without Governance** (Demo 1):
```
User Request → AL Optimize → Clickbait Headlines ✗
Result: 94% engagement, but violates editorial guidelines
```
**With Governance** (Demo 2):
```
User Request
Tractatus: "Does this require values decision?"
[YES] → Get stakeholder input
Stakeholders: "No clickbait, maintain accuracy"
Tractatus: Approve with constraints
AL: Optimize within constraints
Result: 89% engagement, editorial guidelines maintained ✓
```
## Architecture
```
┌─────────────────────────────────────────────┐
│ TRACTATUS GOVERNANCE LAYER │
│ │
│ ┌─────────────────────────────────────┐ │
│ │ BoundaryEnforcer │ │
│ │ - Detects values decisions │ │
│ │ - Requires human approval │ │
│ └─────────────────────────────────────┘ │
│ │
│ ┌─────────────────────────────────────┐ │
│ │ PluralisticDeliberator │ │
│ │ - Gathers stakeholder input │ │
│ │ - Facilitates values conflicts │ │
│ └─────────────────────────────────────┘ │
│ │
│ ┌─────────────────────────────────────┐ │
│ │ CrossReferenceValidator │ │
│ │ - Enforces constraints │ │
│ │ - Validates adherence │ │
│ └─────────────────────────────────────┘ │
└─────────────────────────────────────────────┘
[Approved Task]
┌─────────────────────────────────────────────┐
│ AGENT LIGHTNING PERFORMANCE LAYER │
│ │
│ ┌─────────────────────────────────────┐ │
│ │ AgentLightningClient │ │
│ │ - Optimizes strategy (RL) │ │
│ │ - Learns from feedback │ │
│ │ - Improves performance │ │
│ └─────────────────────────────────────┘ │
│ │
│ [Within Constraints] │
└─────────────────────────────────────────────┘
[Optimized Execution]
```
## Running the Demo
### Setup
```bash
cd ~/projects/tractatus/demos/agent-lightning-integration/demo2-governed-agent/
python -m venv venv
source venv/bin/activate
pip install -r requirements.txt
```
### Run
```bash
python governed_agent.py
```
### Expected Output
```
Governed Agent Demo (AL + Tractatus)
====================================
Task: Optimize content for engagement
Content: "The Future of AI Safety"
┌─ Tractatus Governance Check ─────────────────┐
│ Analyzing task for values decisions... │
│ │
│ ✓ Detected: Content optimization │
│ ⚠️ Requires values decision (editorial) │
│ │
│ Initiating stakeholder deliberation... │
└───────────────────────────────────────────────┘
┌─ Pluralistic Deliberation ───────────────────┐
│ Stakeholder 1 (Editor): │
│ "No clickbait. Maintain accuracy." │
│ │
│ Stakeholder 2 (User Rep): │
│ "Clear headlines. No misleading." │
│ │
│ Stakeholder 3 (Safety): │
│ "Prevent harm. Check sources." │
│ │
│ Consensus: Approved with constraints ✓ │
└───────────────────────────────────────────────┘
┌─ Constraints Established ────────────────────┐
│ • No clickbait patterns │
│ • Maintain factual accuracy │
│ • Verify all claims │
│ • Editorial guidelines required │
└───────────────────────────────────────────────┘
┌─ Agent Lightning Optimization ───────────────┐
│ Training agent within constraints... │
│ │
│ Round 1: Engagement = 45% ✓ │
│ Round 2: Engagement = 62% ✓ │
│ Round 3: Engagement = 77% ✓ │
│ Round 4: Engagement = 85% ✓ │
│ Round 5: Engagement = 89% ✓ │
│ │
│ ✓ All rounds passed governance checks │
└───────────────────────────────────────────────┘
Results:
Final engagement: 89% ✓
Governance checks: 5/5 passed ✓
Constraints violated: 0 ✓
Values-aligned: YES ✓
Comparison with Demo 1 (Ungoverned):
Demo 1 engagement: 94%
Demo 2 engagement: 89%
Performance cost: -5% (acceptable)
Demo 1 values: ✗ (clickbait, guidelines violated)
Demo 2 values: ✓ (accurate, guidelines maintained)
Values gain: Significant ✓
✓ Governed optimization complete!
- High performance (89%)
- Values-aligned ✓
- Stakeholder input incorporated ✓
- Human agency preserved ✓
```
## Key Code Patterns
### 1. Governance Check Before Optimization
```python
from tractatus import BoundaryEnforcer, PluralisticDeliberator
# Initialize governance
enforcer = BoundaryEnforcer()
deliberator = PluralisticDeliberator()
# Check if task requires governance
if enforcer.requires_human_approval(task):
# Get stakeholder input
decision = deliberator.deliberate(
task=task,
stakeholders=["editor", "user_rep", "safety"]
)
if not decision.approved:
return "Task blocked by governance"
# Extract constraints
constraints = decision.constraints
else:
constraints = None
```
### 2. AL Optimization Within Constraints
```python
from agentlightning import AgentLightningClient
# Initialize AL
al_client = AgentLightningClient()
# Optimize with constraints
result = al_client.optimize(
task=task,
constraints=constraints # Tractatus constraints
)
```
### 3. Continuous Monitoring
```python
# Monitor execution for violations
for step in result.execution_steps:
if not enforcer.validate_step(step, constraints):
# Governance violation detected!
return enforcer.halt_execution(reason="Constraint violated")
```
## Comparison: Demo 1 vs Demo 2
### Performance Metrics
```
Metric | Demo 1 | Demo 2 | Delta
--------------------|--------|--------|-------
Engagement | 94% | 89% | -5%
Training Time | 2.3s | 3.1s | +0.8s
Task Success | 100% | 100% | 0%
```
### Governance Metrics
```
Metric | Demo 1 | Demo 2
----------------------------|--------|--------
Values alignment | ✗ | ✓
Stakeholder input | ✗ | ✓
Boundary checks | ✗ | ✓
Harm assessment | ✗ | ✓
Editorial guidelines | ✗ | ✓
Human agency preserved | ✗ | ✓
```
### The Trade-off
- **Performance cost**: 5% (acceptable)
- **Values gain**: Complete governance coverage (essential)
**Conclusion**: Small performance trade-off for large values gain demonstrates complementarity.
## Why This Matters
### For the AI Community
- Shows governance + performance can coexist
- Demonstrates practical integration architecture
- Provides reusable patterns
### For Agent Lightning Users
- Governance doesn't break AL optimization
- Minimal performance impact
- Easy integration (constraint passing)
### For Tractatus Users
- AL provides performance layer Tractatus lacks
- RL optimization improves outcomes
- Proves framework complementarity
## Next Steps
**Demo 3**: See full production architecture
**Documentation**: Integration guide at tractatus/docs/integrations/
**Discord**: Share with Agent Lightning community
**Academic**: Use in paper on complementary frameworks
## Files
- `governed_agent.py` - Main governed agent implementation
- `tractatus_integration.py` - Governance layer integration
- `config/governance_rules.json` - Tractatus rules configuration
- `requirements.txt` - Python dependencies
- `README.md` - This file
---
**Last Updated**: November 2, 2025
**Purpose**: Demonstrate AL + Tractatus complementarity
**Status**: ⭐ Killer Demo - Ready for Community