ultrawork

SKILL.md

Ultrawork

G_meta(lambda, Kappa, Sigma).tau' -- Orchestrate at all scales, compound gains exponentially.

Navigation

Concepts: scale-invariance, metagraph, homoiconicity, pareto-governance, convergence

Integration: delegate-router, learn, ARCHITECTURE

Metasuperhypergraph Architecture

Level 0 (Sigma):     Entire Claude Config (self-referential schema)
Level 1 (G_meta):    Skills + Routers + Agents (hyperedges connecting triads)
Level 2 (G):         Individual component files
Level 3 (V):         Sections within files (frontmatter, body, graph)
Level 4 (v):         Individual properties/concepts

                    SCALE INVARIANCE
    structure(Level_n) ≅ structure(Level_{n+1}) ≅ lambda.o.tau

Core Principles

1. Renormalization Group Optimization

Apply the same optimization at each scale level:

def renormalize(G_level):
    """Coarse-grain, identify universality class, extract relevant operators."""
    # Identify redundant patterns
    redundant = find_duplicates(G_level)
    # Merge into canonical forms
    canonical = merge_to_canonical(redundant)
    # Project to lower-dimension effective theory
    return project(canonical, dim=relevant_operators_only)
Scale Renormalization Action Universality Class
Config Consolidate routers (15 -> 7) Delegation pattern
Skill Merge duplicate content lambda.o.tau triad
Agent Domain agents absorb specialists Execution pattern
File Deduplicate sections Frontmatter schema

2. Power Law Resource Allocation (Pareto)

P(k) ~ k^{-alpha}  where alpha approx 2.5

Top 20% of agents handle 80% of tasks:
  oracle, sisyphus-junior, explore, engineer (4/20 = 20%)

Top 20% of skills deliver 80% of value:
  learn, ultrawork, git-master, lambda-skill, obsidian (5/23 approx 22%)

3. Homoiconic Self-Improvement

# Ultrawork can improve itself
assert ultrawork.can_process(ultrawork.schema) == True

def self_improve(skill):
    analysis = skill.analyze(skill.schema)
    improvements = skill.generate_improvements(analysis)
    validated = skill.validate(improvements)
    return skill.apply(validated) if validated else skill

Agent Routing Matrix (Power-Law Optimized)

Tier 1: High-Frequency (80% of delegations)

Agent Model Complexity Use When
sisyphus-junior Sonnet 0.4-0.7 Focused task execution, implementation
explore Haiku 0.1-0.3 Pattern matching, file discovery, quick search
oracle Opus 0.7-1.0 Root cause analysis, architecture decisions
engineer Sonnet 0.5-0.8 Production-ready implementation

Tier 2: Medium-Frequency (15% of delegations)

Agent Model Complexity Use When
librarian Sonnet 0.3-0.5 Documentation research, codebase understanding
architect Opus 0.8-1.0 System design, multi-domain decomposition
prometheus Sonnet 0.5-0.7 Strategic planning, roadmap design
researcher Sonnet 0.4-0.6 Deep research with MCP tools

Tier 3: Low-Frequency (5% of delegations)

Agent Model Complexity Use When
document-writer Haiku 0.2-0.4 README, API docs, technical writing
multimodal-looker Sonnet 0.3-0.5 Screenshots, diagrams, visual analysis
momus Haiku 0.3-0.5 Critical plan review, devil's advocate
metis Sonnet 0.4-0.6 Pre-planning, hidden requirements
frontend-engineer Sonnet 0.4-0.7 UI/UX, components, styling

External CLI Agents (Token Conservation)

Agent Binary Context Limit Use When
gemini /opt/homebrew/bin/gemini 2M tokens Large context analysis (>100K)
codex ~/.local/bin/codex 128K tokens GPT code generation preference
amp ~/.amp/bin/amp 200K tokens Claude-specific delegation

Execution Patterns

Parallel Execution (Independent Tasks)

# Launch simultaneously in single message
spawn:
  - Task(explore, "find auth files", run_in_background: true)
  - Task(librarian, "search auth docs", run_in_background: true)
  - Task(researcher, "find auth best practices", run_in_background: true)
collect: TaskOutput for each
merge: Deduplicate by content hash

Sequential Execution (Dependent Tasks)

# Chain with explicit dependencies
pipeline:
  - result_1 = Task(explore, "find relevant code")
  - result_2 = Task(oracle, "analyze: ${result_1}")
  - result_3 = Task(engineer, "implement fix: ${result_2}")

Background Execution (Long-Running)

background_operations:
  - Package installation: npm install, pip install, cargo build
  - Build processes: npm run build, make, tsc
  - Test suites: npm test, pytest, cargo test
  - Docker operations: docker build, docker pull
  - Large file operations: >1000 files
  - Subagent delegations: complexity > 0.7

foreground_operations:
  - Quick status: git status, ls, pwd (<5s)
  - File reads/edits
  - Simple commands
  - Verification checks

Convergence Detection

Fixed-Point Termination

def at_fixed_point(state, epsilon=0.001):
    """Detect when further iteration yields no improvement."""
    new_state = iterate(state)
    return distance(state, new_state) < epsilon

def ultrawork_loop(task):
    state = initialize(task)
    while not at_fixed_point(state):
        state = parallel_execute(state)
        state = assess(state)
        if converged(state):
            break
    return finalize(state)

Convergence Thresholds

Pipeline Threshold Use When
R1 (Simple) 0.85 Single-domain, <10 files
R2 (Moderate) 0.92 Multi-domain, 10-50 files
R3 (Complex) 0.96 Architecture-level, >50 files

Verification Checklist (Invariants)

Before stopping, ALL must be true:

  • TODO LIST: Zero pending/in_progress tasks
  • FUNCTIONALITY: All requested features work
  • TESTS: All tests pass (if applicable)
  • ERRORS: Zero unaddressed errors
  • TOPOLOGY: eta >= 4 (if knowledge graph modified)
  • MONOTONICITY: len(K') >= len(K) (no knowledge lost)

If ANY checkbox is unchecked, CONTINUE WORKING.

Integration with Learn Skill

Ultrawork extends learn with:

-- Learn: Sequential knowledge compounding
lambda(o, K, Sigma).tau' = renormalize . compound . assess . execute . route . parse

-- Ultrawork: Parallel orchestration with scale invariance
G_meta(lambda, K, Sigma).tau' = parallelize . delegate . renormalize . compound

Post-Task Learning Loop

after_completion:
  - Extract learnings via learn skill
  - Crystallize patterns with vertex-sharing
  - Update knowledge graph (K -> K')
  - If schema improvement identified: propose to user

Invariants Preserved

Invariant Expression Enforcement
K-monotonicity len(K') >= len(K) Never delete knowledge
Topology eta >= 4 Minimum connectivity maintained
Homoiconicity Sigma.can_process(Sigma) Self-referential capability
Scale Invariance structure(L_n) cong structure(L_{n+1}) Same patterns at all levels
Power Law P(k) ~ k^{-alpha} 80/20 resource allocation

Quick Reference

G_meta(lambda,K,Sigma).tau'   Parallelize -> Delegate -> Renormalize -> Compound
K grows                       Sigma evolves              eta>=4 preserved
Scale-invariant               Power-law optimized        Fixed-point convergent
Weekly Installs
7
GitHub Stars
1
First Seen
Jan 26, 2026
Installed on
codex7
opencode5
claude-code5
antigravity4
github-copilot4
windsurf4