What Is Entropy?
Entropy is often described as "disorder," but it's really about the number of ways a system can be arranged. It explains why time seems to flow in one direction, why your coffee gets cold, and why you can't unscramble an egg.
Explore physics fundamentals →
The Second Law of Thermodynamics
The second law states: In any natural process, total entropy either stays the same or increases. Never decreases.
- Heat flows from hot to cold, never reverse
- Ice melts in warm rooms
- Rooms get messy without effort
- Eggs scramble but don't unscramble
What Entropy Really Means
Microstates and Macrostates
A macrostate is what you observe (temperature, pressure). A microstate is the exact arrangement of particles.
- Many microstates give the same macrostate
- High entropy = many possible microstates
- Low entropy = few possible microstates
- Systems naturally evolve toward states with more possible arrangements
Example: Gas in a Box
- All in one corner: Very few arrangements (low entropy)
- Spread throughout: Vastly more arrangements (high entropy)
- Gas naturally spreads out because there are more ways to be spread
The Arrow of Time
- Physical laws are mostly time-reversible
- But entropy only increases
- The past had lower entropy than the future
- This gives time its "arrow"
The Big Bang started with extremely low entropy—and entropy has been increasing ever since.
Entropy in Daily Life
- There are more broken states than working states
- Disorder is statistically overwhelmingly likely
- Refrigerators move heat out, but motors generate more heat
- Total entropy still increases
- Living things are highly ordered (low entropy)
- But life increases entropy elsewhere (heat, waste)
- Total entropy still increases
Entropy and Information
- High entropy = high randomness = more information to specify state
- Low entropy = orderly = less information needed
- Erasing information generates heat (Landauer's principle)
The Heat Death of the Universe
- Stars burn out
- Energy spreads uniformly
- No temperature differences = no work possible
- Maximum entropy = "heat death"
- Trillions of years away
Entropy is deep—it connects thermodynamics, information, and the nature of time itself.