Wednesday, September 17, 2014

On Entropy

               One of the troubles I find with science education in society is thinking we understand things because we've oversimplified them. This is one of my biggest qualms with most things I hear from non-scientists or even a good number of physical science teachers. One of the biggest oversimplifications I hear is the idea that entropy=disorder. I can’t remember how many dozens of times I have heard this and then seen people take it and run with it. I've even seen some chemistry teachers post videos on YouTube preaching this idea in the completely wrong context. I have seen a quite a few videos lately as well as this gem that has been floating around facebook. While entropy does have to do with disorder in a sense of the word the common understanding of disorder (like a messy room) has nothing to do with entropy. So, I’ve decided to clear up the misconception and attempt to prevent people from associating the word entropy with disorder. If you get lost in the middle you can go ahead and skip to the bottom few paragraphs since I’m sure there is no one that’s actually dedicated enough to read this whole post (if you understand entropy you have no need, if you don’t you probably have no desire).
                First let me explain the simple example of disorder that relates to entropy (the one that I’ve heard most). Think of a room full of nicely organized things, then over time you use the stuff without putting it back. You wear your socks and throw them on the floor, the desk has pens and pencils randomly laying everywhere, and your bed is no longer nicely made. Many people would call this a high state of entropy because of the level of disorder of things in the room. Since the stuff is not neatly organized with must have high entropy. This actually has nothing to do with the actual definition of entropy. Mathematically (which is really the best way to describe it) the entropy of a system is a measure of how many microstates are possible within a single macrostate. For those of us who don’t speak geek I will explain. A macrostate describes the contents of the system and a microstate describes how those contents are arranged. Let me give an example so it makes sense. If I have 6 coins on a table there would be 6 possible macrostates: 6 coins heads up, 5 coins heads up and 1 coin tails up, 4 coins heads up and 2 coins tails up, 3 coins heads up and 3 coins tails up, etc… The microstate would then describe which of the coins would be heads up and tails up. Now that we have that all covered let’s talk about the entropy of this system.
                Like I said before entropy is a measure of how many possible microstates there are. High entropy correlates to a high number of microstates, and low entropy is a low number of microstates. To make things easier to follow I’ll take the number of coins down to four and let’s say they’re all different (a penny, a nickel, a dime, and a quarter). Let’s say I’m OCD and I like to have all my coins facing heads up on the table. I have created a system with very low entropy statistically. That is because there is only one way to arrange the macrostate (all coins heads up) which is to have all the coins heads up. So therefore the number of possible microstates in this macrostate is one. Now lets say I have a really noisy neighbor who has a huge subwoofer which shakes my table while I’m gone. The coins will now start flipping on the table all day. If I were to make any bets on how the coins would be arranged I would say that there would be two coins heads up and two coins heads down. Why? one might ask. It is because two coins heads up is the macrostate which has the most number of possible microstates which means that it has the highest level of entropy in the statistical sense of the word. Let’s look at it closer. I’ve made a table to make it easier to follow. You should note that there are 6 different ways of organizing these coins with two coins heads up and two coins tails up. That’s 5 more microstates or ways of arranging the coins than having them up heads up or all tails up! So while I would know to bet on the macrostate, I would also know not to bet on the microstate as my odds would be lower.

Penny
Nickel
Dime
Quarter
Heads
Heads
Tails
Tails
Heads
Tails
Heads
Tails
Heads
Tails
Tails
Heads
Tails
Heads
Heads
Tails
Tails
Heads
Tails
Heads
Tails
Tails
Heads
Heads

Just to convince that this is the highest statistical entropy state let’s look at the number of ways you can arrange the coins with three coins heads up. The number of possible microstates of this macrostate of the system is 4. That’s still more ways to arrange the coins than all heads up, but less than two heads up and two tails up.

Penny
Dime
Nickel
Quarter
Heads
Heads
Heads
Tails
Heads
Tails
Heads
Heads
Heads
Heads
Tails
Heads
Tails
Heads
Heads
Heads

This phenomenon is also seen in the statistics of flipping a single coin many times. If you flip a coin fifty times you should notice that on average you will get 25 tails and 25 heads. The macrostate would be flipping a coin 50 times and getting 25 tails and 25 heads. The microstate would be the order in which you got the 25 tails and 25 heads. If one were to flip 25 heads in a row it wouldn’t be defying statistics because each flip has a 50% chance of either heads or tails and the previous flip has no affect on the next flip statistically. It would be a rare thing to see however because it would be one of about 126 trillion different ways of obtaining a 50% heads to tails ratio from 50 flips. While it is true that this coin flip example does not happen exactly 50% heads and tails 100% of the time, if one were to take an average of all the unbiased coin flips in the world it should average to very close to that (to the point of being in practical language exactly 50%). Now this is just a statistical explanation of entropy, how does this apply in the real world?
Well it is easiest to see in the sense of an ideal gas system. This is a system made of gas particles that act like billiard balls in a container that see no gravity, no air resistance, no friction, and the collisions are purely elastic (the system does not lose any energy in the collisions). Let’s say I have all my billiard balls arranged so that I have slow moving balls on one side of the container and fast moving balls on the other side of the container with some kind of magical barrier in between. Now before we continue with this box example I can tell you that on average all the balls on one side will have the same average speed. How? Because that is the state with the highest entropy, and the second law of thermodynamics states that a system will always increase in entropy until it reaches maximum entropy. This maximum entropy is often referred to as thermodynamic equilibrium. When one ball collides with another it will transfer some of its kinetic energy to the other ball. After a long time all of the balls will have collided with each other multiple times each. Each collision causes each ball to essentially share its kinetic energy with the other balls. Eventually they will have come to a point where they all have the same average kinetic energy. If this is true than there are many ways of arranging this system since I can say that any ball can be put in any other balls place. If only one ball had all the kinetic energy then that ball is the only ball that I could rearrange which limits my number of microstates.
So now we go back to the box with the magical barrier. Let’s say that there are x number of microstates for each side of the box. That means there are 2 times x number of total microstates in the box since the only way I can rearrange this system is by either putting the fast balls on the right or the slow balls on the right. Now I take the barrier out, what happens? The fast balls start colliding with the slow balls and after a certain amount of time I have all the balls with the same average speed again. That average speed is faster than the average speed of the original slow balls, but slower than the average speed of the original fast balls. That means that the balls could all be rearranged to replace any other ball in the box which correlates to high entropy. Now why would we call this disorder? It is because this system we know very little about. All I can tell you is the average speed of each ball in the box. Before I took the barrier away I was able to tell you a little about the two different average speeds (the slow and the fast).
Now let’s say I rewind time even further and say initially I had only one ball that was moving on each side of the barrier. Well then I can tell you that only two balls are moving at some exact speed, and the rest of the balls are at rest. You see initially I had a highly “ordered” system because I had only two balls with any kinetic energy and I knew exactly what their energy was. After some time the balls bounced around and collided with other balls causing my information about the system to be even vaguer. I do not know anything about any one particular ball now, but I can split the box in two and know that in each half the balls will have some average speed. Now when I take away the barrier I know even less about the system, only that all the balls now have one average speed. One could conclude then that as entropy increases the knowledge one has about a particular system decreases.
This doesn’t mean that 1010 is a low entropy organization of ones and zeros because it repeats a pattern (and therefore probably contains some good information). It merely means that if I had four binary digits that were changing randomly then after some time I would end up with a collection of arrangements of two 1s and two 0s (which includes 1010 as well as 0101, 0011, 1100, and 1001). That means that I know have to guess between 6 options now rather than if they were all 1s I would only have to guess one out of one option. So when you are trying to get rid of information on a hard drive there a repeated pattern of 10 is just as worthless to the NSA as all 1s and then all 0s which is just as worthless as a random assortment of 1s and 0s which are half 1s and half 0s. The information isn’t lost in the actual arrangement; it is lost in the number of possible arrangements. So if there are random 1s and 0s left on your hard drive, but there are more 1s than 0s it makes the NSA’s job easier because it limits the possible combinations of 1s and 0s that contained your actual information previously.
But I digress, now I will get to the whole point of this explanation. Many people say that evolution violates the second law of thermodynamics since our bodies are highly ordered organisms. That would kind of be like saying that because you are using your computer right now you are cheating the laws of physics. In order to get this point across I’m going to take in terms of mass-energy. General relativity states that mass can be converted into energy and vice-versa. This happens all the time not just in fusion in the sun and fission in power plants but in chemical reactions. You see our bodies are highly ordered mass-energy systems. But in order to make our bodies (and keep them running) we actually cause more disorder. For example we eat things which are highly ordered (plant and animal matter or even synthesized chemicals) and we use up the energy that is provided by the chemical bonds of this “fuel”. Well over time we get hungry again. That is because we used this energy and a good portion of it was lost from our body leaking body heat. That energy was partially radiated as electromagnetic radiation and partially lost due to convection in the air or conduction in your clothes/blankets/couch/whatever you are touching. That energy that left you is now “simple” or “less ordered”. This is just like our nuclear fuels which go from high energy density (“high order”) to lower energy density (“low order”). It’s the same with fossil fuels and basically anything that could be rearranged into a lower energy state. This is what is theorized by some to be the “heat death of the universe” or when the entire universe will reach thermodynamic equilibrium. All the higher ordered mass-energy of the universe could eventually all become the same low-ordered mass energy in some form. This would be like everything in the universe evaporating into radio-waves. While no one knows how the universe will end (if it ever “ends”) this goes to show that evolution has no impact on the second law. Yes it is creating “higher order” out of “lower order”, but the total “high order” being converted to “low order” was a net loss in the end.

In conclusion: entropy isn't what most people think it is, and is a very abstract thing that is hard to understand. The one thing I've learned about physics: if you think you know what you’re talking about, you probably don’t.