Chapter 5: The First Law of Complexodynamics
“Interesting things happen on the way from order to chaos.”
Based on: “The First Law of Complexodynamics” (Scott Aaronson, Blog Post)
📄 Original Post: Shtetl-Optimized Blog
5.1 From Physics to Philosophy
Chapter 4 gave us the mathematical machinery—the coffee automaton and sophistication measures. Now we step back to explore the philosophical implications of these ideas.
Scott Aaronson’s blog post “The First Law of Complexodynamics” asks a deceptively simple question:
Why is the universe interesting?
This isn’t mysticism—it’s a precise scientific question with surprising answers.
graph TB
subgraph "The Question"
Q["Why isn't the universe<br/>just boring equilibrium?"]
end
subgraph "Possible Answers"
A1["It's designed that way?<br/>(teleological)"]
A2["We wouldn't be here<br/>to ask otherwise?<br/>(anthropic)"]
A3["Interestingness is<br/>INEVITABLE?<br/>(complexodynamic)"]
end
Q --> A1
Q --> A2
Q --> A3
style A3 fill:#ffe66d,color:#000
Figure: The fundamental question: why isn’t the universe just boring equilibrium? Possible answers include teleological (designed), anthropic (we wouldn’t be here otherwise), or complexodynamic (interestingness is inevitable).
5.2 Stating the First Law
The Law
Aaronson proposes what he calls (somewhat tongue-in-cheek) the First Law of Complexodynamics:
“If a system starts in a low-entropy state and evolves according to simple, reversible dynamics, then its apparent complexity will increase before it decreases.”
Or more simply:
“On the road from simple to random, you must pass through interesting.”
graph LR
S["SIMPLE<br/>(low entropy,<br/>low complexity)"]
I["INTERESTING<br/>(medium entropy,<br/>high complexity)"]
R["RANDOM<br/>(high entropy,<br/>low complexity)"]
S -->|"inevitable"| I
I -->|"inevitable"| R
style I fill:#ffe66d,color:#000
Figure: The First Law of Complexodynamics: systems must pass through an “interesting” phase (medium entropy, high complexity) when evolving from simple (low entropy, low complexity) to random (high entropy, low complexity).
5.3 What Makes Something “Interesting”?
The Goldilocks Zone
Aaronson argues that “interestingness” lives in a Goldilocks zone between order and chaos:
graph TB
subgraph "Too Simple"
TS["All zeros: 0000000000<br/>Perfectly ordered crystal<br/>Blank canvas"]
TSP["Easy to describe,<br/>nothing to discover"]
end
subgraph "Just Right"
JR["Natural language text<br/>Living organism<br/>Working program"]
JRP["Rich structure,<br/>patterns to find"]
end
subgraph "Too Random"
TR["Random noise: 0110100111<br/>Thermal equilibrium<br/>White noise"]
TRP["Nothing to describe,<br/>no patterns"]
end
TS --> TSP
JR --> JRP
TR --> TRP
style JRP fill:#ffe66d,color:#000
Figure: The Goldilocks zone of interestingness. Too simple (all zeros, perfect crystal) is easy to describe but has nothing to discover. Just right (natural language, living organisms) has rich structure and patterns. Too random (noise, equilibrium) has no patterns to describe.
Formal Characterization
Interestingness requires:
- Non-trivial patterns (high sophistication)
- Compressibility (structure exists)
- Not too compressible (not trivial)
5.4 The Inevitability Argument
Why Must Complexity Rise?
Here’s the key insight:
graph TB
subgraph "The Argument"
P1["Start with simple state<br/>(describable in few bits)"]
P2["Evolve with simple rules<br/>(reversible dynamics)"]
P3["End at random-looking state<br/>(thermodynamic equilibrium)"]
end
P1 --> P2 --> P3
subgraph "The Constraint"
C["Reversibility means<br/>information is CONSERVED"]
end
P2 --> C
subgraph "The Conclusion"
CON["The 'structure' in the initial<br/>state must go SOMEWHERE<br/>before disappearing into noise"]
end
C --> CON
style CON fill:#ffe66d,color:#000
Figure: The inevitability argument. Starting from a simple state (describable in few bits), evolving with simple reversible rules, and ending at random-looking equilibrium. Reversibility means information is conserved, so the structure in the initial state must transform into intermediate complexity before disappearing.
The Conservation of Information
Because physics is reversible:
- You can always (in principle) run the movie backward
- Information is never truly destroyed, just spread out
- The initial “simplicity” must transform into intermediate “complexity”
5.5 Why Isn’t Everything Already Boring?
The Deep Question
If entropy always increases and complexity eventually falls, why isn’t the universe already at thermal equilibrium?
graph TB
subgraph "Two Answers"
A1["The universe is young<br/>(hasn't had time)"]
A2["The universe started<br/>in a special state<br/>(low entropy Big Bang)"]
end
Q["Why isn't everything<br/>boring yet?"]
Q --> A1
Q --> A2
BB["The Past Hypothesis:<br/>The universe started<br/>in an extremely<br/>low-entropy state"]
A2 --> BB
style BB fill:#ff6b6b,color:#fff
Figure: Two possible answers to why the universe isn’t boring yet. The anthropic answer: the universe is young and hasn’t reached equilibrium. The complexodynamic answer: the universe started in a special low-entropy state (the Past Hypothesis), which is the source of all structure.
The Past Hypothesis
This is one of the deepest mysteries in physics:
- The universe began in an extraordinarily unlikely state
- This initial “ordered” state is the source of ALL structure we see
- The Second Law is really about evolution FROM that special beginning
5.6 Complexity vs. Entropy: A Deeper Look
Different Curves, Different Meanings
xychart-beta
title "Entropy vs Various Complexity Measures"
x-axis "Time" [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10]
y-axis "Value" 0 --> 100
line "Entropy (always rises)" [5, 20, 35, 50, 62, 73, 82, 89, 94, 97, 99]
line "Kolmogorov Complexity" [10, 35, 55, 70, 80, 87, 92, 95, 97, 98, 99]
line "Sophistication" [5, 30, 55, 70, 65, 50, 35, 22, 12, 6, 2]
line "Logical Depth" [2, 15, 40, 65, 80, 85, 82, 70, 50, 30, 15]
Figure: Entropy vs various complexity measures over time. Entropy monotonically increases, Kolmogorov complexity increases to maximum, while sophistication and logical depth rise then fall. This demonstrates that different measures capture different aspects of complexity.
| Measure | Behavior | What It Captures |
|---|---|---|
| Entropy | Monotonic increase | Disorder, uncertainty |
| Kolmogorov Complexity | Increases to max | Incompressibility |
| Sophistication | Rise then fall | Meaningful structure |
| Logical Depth | Rise then fall | Computational history |
5.7 Aaronson’s Philosophical Insights
On the Nature of Interestingness
Aaronson makes several profound observations:
1. Interestingness is not subjective
graph LR
subgraph "Common View"
CV["'Interesting' is in<br/>the eye of the beholder"]
end
subgraph "Aaronson's View"
AV["'Interesting' has an<br/>objective component:<br/>HIGH SOPHISTICATION"]
end
CV -->|"challenged by"| AV
Figure: Contrasting views of interestingness. The common view sees it as subjective (“in the eye of the beholder”), while Aaronson’s view provides an objective component based on high sophistication—measurable structure in the pattern.
2. We are complexity appreciators
“We evolved to notice patterns, to find compressed descriptions. We ARE the universe looking at itself and finding structure.”
3. The universe is self-sampling
“Any observer will necessarily find themselves in a complex region—simple regions have no observers to appreciate them.”
5.8 Implications for AI and Intelligence
Intelligence as Sophistication Detection
graph TB
subgraph "What Intelligence Does"
I["Intelligence"]
F["Finds patterns<br/>(compression)"]
B["Builds models<br/>(short descriptions)"]
P["Predicts<br/>(uses structure)"]
end
I --> F
I --> B
I --> P
C["Intelligence is a<br/>SOPHISTICATION DETECTOR"]
F --> C
B --> C
P --> C
style C fill:#ffe66d,color:#000
Figure: What intelligence does: finds patterns (compression), builds models (short descriptions), and predicts (uses structure). Intelligence is fundamentally a sophistication detector—it identifies and exploits meaningful structure in data.
For Machine Learning
The implications for ML are profound:
- Learnable data has high sophistication
- Random data → nothing to learn
- Simple data → trivial to learn
- Interesting data → the sweet spot
- Learning = finding the sophisticated part
- Neural nets extract structure from noise
- The learned representation IS the sophistication
- Generalization requires sophistication
- If the test set has the same structure as training
- The model can generalize
- Random test data → no generalization possible
5.9 The Computational Lens
Complexity Requires Computation
Aaronson emphasizes a computational perspective:
Complex structures are “frozen accidents” of computation.
graph TB
subgraph "Creating Complexity"
I["Initial conditions<br/>+ Simple rules"]
C["Computation<br/>(time evolution)"]
O["Complex output<br/>(high sophistication)"]
end
I --> C --> O
D["The complexity is<br/>in the PROCESS,<br/>not just the state"]
C --> D
style D fill:#ffe66d,color:#000
Figure: Creating complexity requires initial conditions plus simple rules, which through computation (time evolution) produce complex output with high sophistication. The complexity is in the process itself, not just the final state.
Logical Depth
Logical depth (from Charles Bennett) measures:
How much computation is needed to produce a state from its shortest description?
- Random strings: SHORT depth (just output random bits)
- Simple patterns: SHORT depth (trivial to generate)
- Interesting structures: LONG depth (required computation)
5.10 The Broader Picture
A New Framework for Understanding Reality
graph TB
subgraph "Traditional Physics"
TP["Focus on<br/>fundamental laws"]
end
subgraph "Thermodynamics"
TD["Focus on<br/>entropy & equilibrium"]
end
subgraph "Complexodynamics"
CD["Focus on<br/>emergence of structure"]
end
TP --> TD --> CD
Q["Answers different question:<br/>Not 'what are the laws?'<br/>Not 'where does energy go?'<br/>But 'why do interesting<br/>things happen?'"]
CD --> Q
style Q fill:#ffe66d,color:#000
Figure: Complexodynamics provides a new framework for understanding reality, complementing traditional physics (fundamental laws) and thermodynamics (entropy). It answers a different question: why do interesting things happen, focusing on the emergence of structure.
The Three Regimes
| Regime | Entropy | Complexity | Examples |
|---|---|---|---|
| Ordered | Low | Low | Crystal, empty space |
| Complex | Medium | High | Life, galaxies, brains |
| Chaotic | High | Low | Thermal equilibrium, noise |
5.11 Criticisms and Limitations
What the Theory Doesn’t Explain
Aaronson is careful to note limitations:
graph TB
subgraph "Doesn't Explain"
D1["WHY the universe<br/>started ordered"]
D2["The specific FORM<br/>of complexity"]
D3["Consciousness or<br/>subjective experience"]
D4["Quantitative predictions<br/>for real systems"]
end
subgraph "Does Explain"
E1["Why complexity<br/>is POSSIBLE"]
E2["Why it's<br/>INEVITABLE"]
E3["Why it's<br/>TEMPORARY"]
end
Figure: Limitations of complexodynamics. It doesn’t explain why the universe started ordered, the specific form of complexity, consciousness, or quantitative predictions. However, it does explain why complexity is possible, inevitable, and temporary.
Open Questions
- Can we precisely measure sophistication for real-world systems?
- What’s the relationship between complexity and consciousness?
- How does quantum mechanics affect these arguments?
- Can we engineer systems to maximize complexity duration?
5.12 Connections to Other Chapters
graph TB
CH5["Chapter 5<br/>First Law of<br/>Complexodynamics"]
CH5 --> CH4["Chapter 4: Coffee Automaton<br/><i>The mathematical model</i>"]
CH5 --> CH2["Chapter 2: Kolmogorov<br/><i>Complexity definitions</i>"]
CH5 --> CH1["Chapter 1: MDL<br/><i>Finding structure</i>"]
CH5 --> CH27["Chapter 27: Superintelligence<br/><i>Intelligence emergence</i>"]
CH5 --> CH25["Chapter 25: Scaling Laws<br/><i>Optimal complexity</i>"]
style CH5 fill:#ff6b6b,color:#fff
Figure: The First Law of Complexodynamics connects to multiple chapters: the Coffee Automaton (mathematical model), Kolmogorov complexity (sophistication definitions), MDL (finding structure), superintelligence (intelligence emergence), and scaling laws (optimal complexity).
5.13 Synthesis: What Part I Has Taught Us
The Complete Picture
We can now synthesize all of Part I:
graph TB
subgraph "Part I: Foundations"
CH1["Ch 1: MDL<br/>Best model = shortest description"]
CH2["Ch 2: Kolmogorov<br/>Complexity = program length"]
CH3["Ch 3: NN Simple<br/>Training = compression"]
CH4["Ch 4: Coffee<br/>Complexity rises & falls"]
CH5["Ch 5: This Chapter<br/>Interestingness is inevitable"]
end
CH1 --> S["SYNTHESIS"]
CH2 --> S
CH3 --> S
CH4 --> S
CH5 --> S
S --> C["Intelligence finds patterns.<br/>Patterns are inevitable.<br/>The universe creates<br/>its own observers."]
style C fill:#ffe66d,color:#000,stroke:#000,stroke-width:2px
Figure: Synthesis of Part I foundations. MDL provides the framework (minimize description length), Kolmogorov complexity provides the measure (shortest program), keeping NNs simple applies it (regularization), the Coffee Automaton shows the dynamics (complexity rises and falls), and Complexodynamics provides the philosophy (interestingness is inevitable).
The Meta-Lesson
The same principles that govern cream mixing in coffee govern the emergence of intelligence in the universe—and the training of neural networks on your laptop.
5.14 Key Equations Summary
The First Law (Informal)
\(\text{Simple} \xrightarrow{\text{time}} \text{Complex} \xrightarrow{\text{time}} \text{Random}\)
Sophistication (Reminder)
\(\text{soph}(x) = \min\{K(S) : x \in S, \text{ S is a "pattern"}\}\)
Interestingness Condition
\(\text{Interesting} \Leftrightarrow \text{soph}(x) \gg 0 \text{ AND } K(x) - \text{soph}(x) \gg 0\)
The Information Conservation
\(I(\text{past} : \text{future}) = \text{constant}\) (In reversible dynamics, mutual information is conserved)
5.15 Chapter Summary
graph TB
subgraph "Key Takeaways"
T1["'Interesting' has an<br/>objective definition:<br/>high sophistication"]
T2["Complexity rising then<br/>falling is INEVITABLE,<br/>not accidental"]
T3["Intelligence evolved to<br/>detect sophistication—<br/>we are pattern finders"]
T4["The universe isn't<br/>designed to be interesting—<br/>it CAN'T HELP but be"]
end
T1 --> C["The First Law of Complexodynamics:<br/>On the road from order to chaos,<br/>you must pass through interesting.<br/>We are the universe's way of<br/>noticing itself."]
T2 --> C
T3 --> C
T4 --> C
style C fill:#ffe66d,color:#000,stroke:#000,stroke-width:2px
Figure: Key takeaways from the First Law of Complexodynamics: interestingness is inevitable (not just possible), it’s objective (measurable via sophistication), it’s temporary (rises then falls), and it enables intelligence (which thrives at the complexity peak).
In One Sentence
The First Law of Complexodynamics tells us that interesting, complex structures are not accidental but inevitable—any system evolving from order toward disorder must pass through a phase of high complexity, explaining why the universe produces galaxies, life, and minds.
🎉 Part I Complete!
You’ve finished the Foundations section. You now understand:
- How to measure the quality of explanations (MDL)
- What complexity really means (Kolmogorov)
- How neural networks relate to compression (Hinton)
- Why complexity rises and falls (Coffee Automaton)
- Why interestingness is inevitable (Complexodynamics)
Next up: Part II - Convolutional Neural Networks, where we apply these theoretical foundations to the practical revolution in computer vision.
Exercises
-
Reflection: In your own words, explain why “random” and “simple” are both “uninteresting” while structured patterns in between are “interesting.”
-
Application: Consider a language model like GPT. How does the First Law of Complexodynamics relate to what these models learn?
-
Critique: What are the weaknesses of defining “interestingness” in terms of sophistication? Can you think of things that are intuitively interesting but might have low sophistication?
-
Speculation: If the universe is heading toward heat death (maximum entropy, zero complexity), what does this imply for the very long-term future of intelligence?
References & Further Reading
| Resource | Link |
|---|---|
| Original Blog Post (Aaronson) | Shtetl-Optimized |
| Coffee Automaton Paper | arXiv:1405.6903 |
| Logical Depth (Bennett) | Springer |
| The Past Hypothesis (Carroll) | arXiv:physics/0210022 |
| Sophistication & Depth (Antunes et al.) | Paper |
| Complexity in Physics (Crutchfield) | arXiv:nlin/0508006 |
| Why Now? (Livio & Rees) | arXiv:astro-ph/0503166 |
Next Chapter: Chapter 6: AlexNet - The ImageNet Breakthrough — We begin Part II by exploring the paper that ignited the deep learning revolution: Krizhevsky, Sutskever, and Hinton’s game-changing CNN that dominated ImageNet 2012.