What is Entropy?
“Entropy is the arrow of time, the witness of the universe’s irreversibility, the measure from order to chaos—but it is far deeper than you imagine.”
← Previous: What is Scattering | Back to Home | Next: Foundation Summary →
Starting from Room Chaos
Imagine your room:
Morning (just tidied):
- Books neatly arranged on shelves
- Clothes folded in wardrobe
- Desk clean and tidy
Evening (after a day):
- Books scattered on table and bed
- Clothes piled on chair
- Desk in complete disarray
graph LR
Morning["Morning<br/>Ordered<br/>Low Entropy"] -->|time| Evening["Evening<br/>Chaotic<br/>High Entropy"]
Evening -.very hard.-> Morning
style Morning fill:#a8e6cf
style Evening fill:#ffaaa5
Question: Why does the room always get messier? Why doesn’t it tidy itself?
Answer: Entropy
Definition of Entropy: Measure of Chaos
📊 Statistical Definition (Boltzmann)
Entropy is a measure of the system’s “chaos” or “number of possible microstates.”
Boltzmann formula:
where:
- = entropy
- = Boltzmann constant
- = number of microstates (how many ways to realize this macrostate)
Example: Playing cards
graph TB
Ordered["Ordered Arrangement<br/>(by suit and rank)<br/>Ω = 1<br/>S = 0"] -->|shuffle| Random["Random Arrangement<br/>Ω = 52!≈8×10^67<br/>S = k_B ln(52!)"]
style Ordered fill:#a8e6cf
style Random fill:#ffaaa5
-
Ordered arrangement: Only one way (A♠, 2♠, …, K♠, A♥, …)
- ,
-
Random arrangement: There are ways
- , (huge)
💡 Key Insight: High entropy = chaos = many possibilities; Low entropy = order = few possibilities
🌡️ Thermodynamic Definition (Clausius)
In thermodynamics, entropy is defined as:
where:
- = change in entropy
- = heat transfer in reversible process
- = temperature
Physical meaning:
- Entropy measures “unavailable energy”
- Lower temperature means same heat corresponds to larger entropy change
- Irreversible processes produce entropy
Second Law of Thermodynamics: Entropy Always Increases
📈 The Most Important Law of the Universe
Second Law of Thermodynamics:
In an isolated system, entropy always increases or remains constant, never decreases.
graph LR
Past["Past<br/>Low Entropy<br/>Ordered"] -->|time| Future["Future<br/>High Entropy<br/>Chaotic"]
Future -.impossible.-> Past
style Past fill:#a8e6cf
style Future fill:#ffaaa5
Examples:
- Ice melting: Ordered crystal → Disordered water (entropy increase)
- Perfume diffusion: Concentrated perfume molecules → Uniform distribution (entropy increase)
- Breaking an egg: Intact shell → Fragments (entropy increase)
- Universe expansion: Dense Big Bang → Sparse galaxies (entropy increase)
⏰ Entropy and Time Arrow
Why do we remember the past but not the future? Why does time have a direction?
Answer: Because entropy is increasing!
graph TD
Arrow["Time Arrow"] --> Thermo["Thermodynamic Arrow<br/>Entropy Always Increases"]
Arrow --> Causal["Causal Arrow<br/>Cause before Effect"]
Arrow --> Psych["Psychological Arrow<br/>Remember Past, Not Future"]
Thermo -.determines.-> Causal
Causal -.determines.-> Psych
style Arrow fill:#ff6b6b,stroke:#c92a2a,stroke-width:3px,color:#fff
Three arrows, one essence:
- Thermodynamic arrow: Direction of entropy increase
- Causal arrow: Cause before effect
- Psychological arrow: Direction of memory
They all point in the same direction—the direction of entropy increase!
Information Entropy: Measure of Surprise
📡 Shannon Entropy
In information theory, entropy measures “information content” or “uncertainty.”
Shannon entropy formula:
where:
- = information entropy
- = probability of event occurring
Example: Coin toss
graph TB
Fair["Fair Coin<br/>p(heads)=0.5, p(tails)=0.5"] -->|entropy| H1["H = -0.5 ln 0.5 - 0.5 ln 0.5<br/>= ln 2 ≈ 0.693 bit"]
Biased["Biased Coin<br/>p(heads)=0.9, p(tails)=0.1"] -->|entropy| H2["H = -0.9 ln 0.9 - 0.1 ln 0.1<br/>≈ 0.325 bit"]
style H1 fill:#ffe66d,stroke:#f59f00,stroke-width:2px
style H2 fill:#e0e0e0
- Fair coin: Most uncertain ( maximum)
- Biased coin: Relatively certain ( smaller)
- Deterministic outcome: Completely certain ()
💡 Key Insight: Entropy = degree of surprise. The more uncertain something is, the more surprising when it occurs, the greater the entropy.
🔗 Information Entropy = Thermodynamic Entropy
Remarkably, it is widely accepted in physics that information entropy and thermodynamic entropy have a profound connection!
Landauer’s principle:
Erasing 1 bit of information requires dissipating at least of energy, producing of entropy.
This shows: Information is physical!
Relative Entropy: Measure of Distance
📏 Kullback-Leibler Divergence
Relative entropy (KL divergence) measures the “distance” between two probability distributions:
Or continuous version:
Properties:
- Non-negativity:
- Asymmetry:
- Monotonicity: Monotonically decreasing under certain evolutions
graph LR
Rho["Distribution ρ"] -.relative entropy.-> Sigma["Distribution σ"]
Rho -.distance=0.-> Rho2["ρ = σ"]
style Rho fill:#ffd3b6
style Sigma fill:#a8e6cf
style Rho2 fill:#ffe66d,stroke:#f59f00,stroke-width:2px
Physical meaning:
In GLS theory, the monotonicity of relative entropy is considered the foundation of the time arrow!
The system always evolves toward equilibrium, relative entropy monotonically decreasing.
Generalized Entropy: Area + Matter
🕳️ Generalized Entropy of Black Holes
In gravitational systems, entropy includes not only matter entropy but also geometric entropy:
graph TB
Gen["Generalized Entropy<br/>S_gen"] --> Area["Geometric Entropy<br/>A/4Gℏ<br/>(Black Hole Horizon Area)"]
Gen --> Matter["Matter Entropy<br/>S_out<br/>(Matter Outside Horizon)"]
style Gen fill:#ff6b6b,stroke:#c92a2a,stroke-width:3px,color:#fff
Bekenstein-Hawking entropy:
Black hole entropy is proportional to horizon area:
Example: Solar mass black hole
- Mass: kg
- Schwarzschild radius: km
- Area: m²
- Entropy:
This is enormous! Much larger than the entropy of gas of the same mass.
📊 Generalized Second Law
Generalized Second Law (GSL):
Generalized entropy always increases or remains constant.
Hawking’s thought experiment:
Throw a book into a black hole:
- Book falls into black hole → External matter entropy decreases ()
- Black hole mass increases → Horizon area increases ()
- Total effect: (generalized entropy still increases)
graph LR
Book["Book<br/>Mass m<br/>Entropy S_book"] -->|falls into| BH["Black Hole<br/>Mass M→M+m"]
BH --> Result["Result<br/>ΔS_out = -S_book<br/>ΔA/4G > S_book<br/>ΔS_gen > 0"]
style BH fill:#000,color:#fff
style Result fill:#a8e6cf
Entropy and Causality: Unification in GLS Theory
In GLS unified theory, entropy plays a central role:
🔗 Causality = Entropy Monotonicity
Remember what we said in “What is Causality”?
GLS theory infers: Causal order is mathematically equivalent to entropy monotonicity!
This means:
- Saying “A is before B” = saying “A’s entropy ≤ B’s entropy”
- Time arrow = direction of entropy increase
- Causality = partial order relation of entropy
graph TD
Causality["Causality<br/>A ≺ B"] -.equivalent.-> EntropyMonotone["Entropy Monotonicity<br/>S(A) ≤ S(B)"]
EntropyMonotone -.equivalent.-> TimeOrder["Time Order<br/>t(A) ≤ t(B)"]
TimeOrder -.equivalent.-> Causality
style Causality fill:#ff6b6b,stroke:#c92a2a,stroke-width:3px,color:#fff
📐 Entropy Extremum on Small Causal Diamonds
One of the core insights of GLS theory:
GLS theory derivation shows: On small causal diamonds, generalized entropy takes an extremum if and only if Einstein’s equation holds.
Information Geometric Variational Principle (IGVP):
In plain language:
- On small causal diamonds, require generalized entropy to take an extremum
- This extremum condition automatically derives Einstein’s equation!
Significance:
GLS theory suggests: Gravity might not be a fundamental force, but a geometric emergence of entropy extremum!
graph TB
Diamond["Small Causal Diamond"] --> Entropy["Generalized Entropy<br/>S_gen = A/4Gℏ + S_out"]
Entropy --> Extremum["Extremum Condition<br/>δS_gen = 0"]
Extremum --> Einstein["Einstein's Equation<br/>G_ab + Λg_ab = 8πGT_ab"]
style Entropy fill:#ffe66d,stroke:#f59f00,stroke-width:2px
style Einstein fill:#4ecdc4,color:#fff
Just as soap bubbles automatically form spheres (minimum surface area), spacetime automatically satisfies Einstein’s equation (entropy extremum)!
Deeper Meaning of Entropy
🌌 Ultimate Fate of the Universe
If entropy always increases, what is the final state of the universe?
Heat death:
- All energy uniformly distributed
- No temperature differences
- No available energy to do work
- Entropy reaches maximum
- Time “stops” (no change)
graph LR
BigBang["Big Bang<br/>Low Entropy<br/>Highly Ordered"] -->|time| Now["Now<br/>Medium Entropy"]
Now -->|time| HeatDeath["Heat Death<br/>Maximum Entropy<br/>Complete Chaos"]
style BigBang fill:#a8e6cf
style Now fill:#ffe66d,stroke:#f59f00,stroke-width:2px
style HeatDeath fill:#e0e0e0
Time scale: About years (far exceeding the universe’s current age of 13.8 billion years)
🤔 Mystery of Low-Entropy Past
If entropy always increases, why was entropy so low at the beginning (Big Bang)?
This is one of the unsolved mysteries of physics!
Possible explanations:
- Cosmological principle: Initial conditions at universe’s beginning were low-entropy (but why?)
- Special nature of gravity: Entropy of gravitational systems differs from other systems
- Multiverse: Our universe is one of many that happens to be low-entropy
- GLS theory: Boundary conditions might determine initial low entropy
Summary: Multiple Faces of Entropy
| Perspective | What is Entropy | Formula | Analogy |
|---|---|---|---|
| Statistical Mechanics | Number of microstates | Room chaos | |
| Thermodynamics | Unavailable energy | Dissipated energy | |
| Information Theory | Uncertainty | Degree of surprise | |
| Black Hole Physics | Horizon area | Holographic encoding | |
| Relative Entropy | Distribution distance | KL divergence | |
| GLS Theory | Causal order | Time arrow |
🎯 Key Points
- Second Law of Thermodynamics: Entropy always increases (isolated systems)
- Time arrow: Direction of entropy increase is the direction of time
- Information = Physics: Information entropy and thermodynamic entropy are essentially the same
- Generalized entropy:
- Causality = Entropy: Causal order is equivalent to entropy monotonicity
- IGVP: Entropy extremum derives Einstein’s equation
💡 Most Profound Insight
GLS theory proposes: Entropy is not just “chaos,” it might be the arrow of time, the order of causality, the source of gravity. All evolution in the universe is essentially a process of entropy increase.
Entropy unifies thermodynamics, information theory, gravity, and causality:
- Thermodynamics: Entropy = dissipation of energy
- Information theory: Entropy = measure of information
- Gravity: Entropy = area of horizon
- Causality: Entropy = arrow of time
They are all different aspects of the same “entropy”!
Next
Congratulations! You have learned five fundamental concepts: time, causality, boundary, scattering, entropy.
Now it’s time to summarize and see how they fit together into a complete picture:
There, we will see how these five concepts merge into one in GLS unified theory.
Remember: Entropy is one of the most profound concepts in the universe. It not only tells us “why rooms get messy,” but also “why time has direction,” “why gravity exists,” “why the universe evolves.” Understanding entropy, you understand the essence of cosmic change.
← Previous: What is Scattering | Back to Home | Next: Foundation Summary →