Workflow System
The workflow system executes AI skills programmatically and tracks their execution history.
Overview
Skills are invoked via the Claude CLI using stream-json format, which allows proper skill expansion and tool access. The workflow executor:
- Invokes a skill by name
- Captures execution metrics (duration, cost, turns)
- Logs results to this file
- Commits changes using the
/agent-commitskill for meaningful messages
Available Skills
Orchestration
The evolution loop (scripts/evolve_loop.py) is the main orchestrator. It runs a deterministic 24-slot task cycle with time-triggered events like daily highlights at 8am UTC.
| Skill | Purpose | Modifies Content? |
|---|---|---|
/replenish-queue [mode] | Auto-generate tasks when queue is empty or near-empty | Yes (todo.md only) |
/tune-system | Monthly meta-review—analyze system operation, adjust cadences/thresholds | Yes (state, minor) |
Content Creation
| Skill | Purpose | Modifies Content? |
|---|---|---|
/expand-topic [topic] | Generate new article on a topic | Yes (creates draft) |
/refine-draft [file] | Improve existing draft content | Yes (edits content) |
/research-topic [topic] | Web research, outputs notes to research | Research notes only |
/research-voids | Daily research on cognitive gaps and unchartable territories | Research notes only |
Review & Validation
| Skill | Purpose | Modifies Content? |
|---|---|---|
/validate-all | Check frontmatter, links, orphans | No (reports only) |
/check-tenets | Verify alignment with 5 foundational tenets | No (reports only) |
/check-links | Verify all internal links work | No (reports only) |
/pessimistic-review | Find logical gaps, unsupported claims, counterarguments | No (reports only) |
/optimistic-review | Find strengths and expansion opportunities | No (reports only) |
/deep-review [file] | Comprehensive single-document review with improvements | Yes (modifies content) |
Content Maintenance
| Skill | Purpose | Modifies Content? |
|---|---|---|
/coalesce | Merge overlapping articles into unified pieces, archiving originals | Yes (creates, archives) |
Publishing
| Skill | Purpose | Modifies Content? |
|---|---|---|
/add-highlight [topic] | Add item to What’s New page (max 1/day). Supports backlog: can highlight any content not featured in last 90 days | Yes (highlights.md) |
Internal (Automation Only)
| Skill | Purpose | Modifies Content? |
|---|---|---|
/agent-commit | Analyze changes and create meaningful git commit messages | Git only |
The /agent-commit skill is invoked automatically by the evolution loop after each content-modifying skill completes. It:
- Receives the previous skill’s output as context
- Runs
git diffto analyze actual file changes - Generates a descriptive commit message (e.g.,
refine(deep-review): improve clarity in free-will.md) - Creates the commit with agent authorship
This replaces the previous generic commit messages like auto(deep-review): Automated execution.
Queue Replenishment
The task queue in todo auto-replenishes when active tasks (P0-P2) drop below 3. The evolution loop triggers /replenish-queue automatically when the queue is low.
Task Types and Chains
Tasks generate follow-up tasks automatically:
| Type | Description | Generates |
|---|---|---|
research-topic | Web research producing notes | → expand-topic |
expand-topic | Write new article | → cross-review |
cross-review | Review article in light of new content | (terminal) |
refine-draft | Improve existing draft | (terminal) |
deep-review | Comprehensive single-doc review | (terminal) |
Task Generation Sources
/replenish-queue generates tasks from four sources:
- Task chains: Recent
research-topiccompletions that need articles written; recentexpand-topiccompletions that need cross-review integration - Unconsumed research: Research notes in
research/without corresponding articles - Gap analysis: Content gaps based on tenet support, undefined concepts, coverage targets
- Staleness: AI-generated content not reviewed in 30+ days
Replenishment Modes
conservative: 3-5 high-confidence tasks only- (default): 5-8 tasks with good diversity
aggressive: 8-12 tasks including speculative ones
Cross-Review Tasks
When a new article is written, /replenish-queue generates cross-review tasks for related existing articles. These reviews:
- Add wikilinks to the new content
- Check for arguments that the new content supports or challenges
- Ensure consistent terminology
- Identify missing cross-references
System Tuning
The /tune-system skill provides meta-level self-improvement for the automation system. It runs monthly (30-day cadence, injected when 45 days overdue).
What It Analyzes
- Cadence adherence: Are maintenance tasks running on schedule or frequently overdue?
- Failure patterns: What’s causing systematic task failures?
- Queue health: Is replenishment producing tasks that actually get executed?
- Review findings: Are identified issues being addressed?
- Convergence progress: Is the system making progress toward goals?
Change Tiers
| Tier | Scope | Approval |
|---|---|---|
| Tier 1 | Cadence ±2 days, threshold ±2 days | Automatic (max 3/session) |
| Tier 2 | New P3 tasks, larger changes | Recommendation only |
| Tier 3 | Skill changes, tenet-related | Report only |
Safeguards
- Evidence threshold: Requires 5+ data points before making changes
- Change cooldown: Settings can’t change twice within 60 days
- Locked settings: Human can lock any setting via
locked_settingsin state - Abort conditions: Stops if >50% failure rate or convergence regresses
Output
Creates report at reviews/system-tune-YYYY-MM-DD.md documenting findings, changes applied, and recommendations.
Running Workflows
Evolution Loop
The evolution loop runs continuously, executing tasks on a 24-slot cycle:
# Run evolution loop (Ctrl+C to stop)
python scripts/evolve_loop.py --interval 2400
# Describe the task cycle
python scripts/evolve_loop.py --describe-cycle
# Test with limited iterations
python scripts/evolve_loop.py --max-iterations 5
Individual Skills
# Run a skill manually
uv run python scripts/run_workflow.py validate-all
# Run with more turns for complex tasks
uv run python scripts/run_workflow.py expand-topic --max-turns 30
Execution Format
Each workflow execution logs:
- Status: Success, Error, MaxTurns, or PermissionDenied
- Duration: How long the execution took
- Cost: API cost in USD
- Turns: Conversation turns used vs maximum
- Output: Brief summary or error message
- Session: Session ID for debugging
Recent Executions
| Title | Created | Modified |
|---|---|---|
Tag Vocabulary This file is the single source of truth for tags. Before adding a tag to any article, check here first.
Rules Use existing tags — never invent a new tag when an existing one covers the concept Lowercase kebab-case only — human-ai-collaboration, never HumanAI or Human AI 3–8 tags per article — prefer specificity over breadth To add a new tag — add it here first with a description and example use, then apply it Synonyms are forbidden — if two tags mean the same thing, retire one and note the canonical form Vocabulary AI & Technology Tag Meaning Do NOT use instead ai-tools General-purpose AI software tools (ChatGPT, Claude, Copilot) used in professional work ai, generative-ai for tool-specific contexts generative-ai Generative AI as a technology category, capability, or paradigm ai (too broad), ai-tools (too specific) ai-agents Autonomous or semi-autonomous AI systems acting on behalf of users automation, agents prompt-engineering The practice of crafting prompts as … | 2026-03-11 | 2026-03-11 |
Highlights Featured content. Updated by the /add-highlight skill.
No highlights yet. Content will appear here as the framework generates articles.
2026-03-11: Steal Like an Artist, Feel Like a Taoist Active curation meets effortless flow. Austin Kleon’s ‘steal’ principle paired with Taoist wu wei shows how bold borrowing and ego-release create the conditions for genuine creativity.
Type: new-article
Link: steal-like-an-artist-feel-like-a-taoist
2026-03-10: Apex Articles: Where Ideas Converge Into Argument Intent Suite is building a synthesis layer—apex articles that weave topics, concepts, and arguments into unified narratives. Each must defend a clear thesis. | 2026-03-09 | 2026-03-11 |
Task Queue Tasks are picked up by the evolution loop (scripts/evolve_loop.py).
Priority Levels P0 — Urgent: blocking issues P1 — High: important improvements P2 — Normal: standard content work (automation picks these) P3 — Low: backlog / nice to have Active Tasks P1: Can Generative AI have an “intent”? Type: research-topic Notes: How to reframe this looking at “intent”? The user has an intent, the why doing something its actions and relations with technology. But can we also flip it? Is there something like artificial intent? Generated: 2026-03-10 ✓ 2026-03-11: How does cognitive debt accumulate in knowledge work that relies heavily on AI? Type: research-topic Notes: Follow-up to AI intensification question. What recovery and boundary practices reduce it? Output: cognitive-debt-ai-knowledge-work-2026-03-11 P2: How is AI reshaping workshop facilitation and design sessions? Type: research-topic Notes: Domain exploration — facilitation. What new roles, risks, and … | 2026-03-10 | 2026-03-10 |
2026-03-11T14:37:00+00:00 - research-topic Status: Success Topic: What does “good enough” mean in AI-augmented systemic design? Output: good-enough-ai-augmented-systemic-design-2026-03-11 Sources consulted: 11 2026-03-11T13:57:00+00:00 - expand-topic Status: Success Topic: The Intent Stack: Making Human Purpose Legible to AI Output: intent-stack-framework Word count: 1166 Based on research: yes — intent-stack-framework-2026-03-11 2026-03-11T13:23:00+00:00 - research-topic Status: Success Topic: The “Expert Benchmark” Fallacy in AI Evaluation Output: expert-benchmark-fallacy-ai-evaluation-2026-03-11 Sources consulted: 8 Changelog AI automation activity log. Updated automatically by the evolution loop. | 2026-03-10 | 2026-03-10 |