Lumifer comments on Entropy and Temperature - Less Wrong Discussion
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (96)
I am not quite sure in which way this statement is useful.
"..and for an encore goes on to prove that black is white and gets himself killed on the next zebra crossing." -- Douglas Adams
I had that thought as well, but the 'Second Law Trickery' section convinced me that it was a useful statement.
I'll grant that it is an interesting statement, but at the moment my impression is that it's just redefining the word "temperature" in a particular way.
Is that because you didn't read the rest of the post?
"Temperature is in the mind" doesn't mean that you can make a cup of water boil just by wishing hard enough. It means that whether or not you should expect a cup of water to boil depends on what you know about it.
(It also doesn't mean that whether an ice cube melts depends on whether anyone's watching. The ice cube does whatever the ice cube does in accordance with its initial conditions and the laws of mechanics.)
So now that you've told me what it does NOT mean, perhaps you can clarify what it DOES mean? I still don't understand.
In particular, the phrase "in the mind" implies that temperature requires a mind and would not exist if there were no minds around. Given that we are talking about classical systems, this seems an unusual position to take.
Another implication of "in the mind" is that different minds would see temperature differently. In fact, if you look into the original EY post, it explicitly says
And that makes me curious about phase changes. Can I freeze water into ice by knowing more about it? Note: not by doing things like separating molecules by energy and ending up with ice and electricity, but purely by knowing?
I don't know of any way that statement in particular is useful, but understanding the model that produces it can be helpful. For example, it's possible to calculate the minimum amount of energy necessary to run a certain computation on a computer at a certain temperature. It's further useful in that it shows that if the computation is reversible, there is no minimum energy.
The model is fine, what I'm having problems with is the whole "in the mind" business which goes straight to philosophy and seems completely unnecessary for the discussion of properties of classic systems in physics.
Entropy is statistical laws. Thus, like statistics, it's in the mind. It's also no more philosophical than statistics is, and not psychological at all.
I have a feeling you're confusing the map and the territory. Just because statistics (defined as a toolbox of methods for dealing with uncertainty) exists in the mind, there is no implication that uncertainty exists only in the mind as well. Half-life of a radioactive element is a statistical "thing" that exists in real life, not in the mind.
In the same way, phase changes of a material exist in the territory. You can usefully define temperature as a particular metric such that water turns into gas at 100 and turns into ice at zero. Granted, this approach has its limits but it does not seem to depend on being "in the mind".
The half-life of a radioactive element is something that can be found without using probability. It is the time it takes for the measure of the universes in which the atom is still whole to be exactly half of the initial measure. Similarly, phase change can be defined without using probability.
The universe may be indeterministic (though I don't think it is), but all this means is that the past is not sufficient to conclude the future. A mind that already knows the future (perhaps because it exists further in the future) would still know the future.
So, does your probability-less half-life require MWI? That's not a good start. What happens if you are unwilling to just assume MWI?
Why do you think such a thing is possible?
Even without references to MWI, I'm pretty sure you can just say the following: if at time t=0 you have an atom of carbon-14, at a later time t>0 you will have a superposition of carbon-14 and nitrogen-14 (with some extra stuff). The half-life is the value of t for which the two coefficients will be equal in absolute value.
Uncertainty in the mind and uncertainty in the territory are related, but they're not the same thing, and calling them both "uncertainty" is misleading. If indeterminism is true, there is an upper limit to how certain someone can reliably be about the future, but someone further in the future can know it with perfect certainty and reliability.
If I ask if the billionth digit of pi is even or odd, most people would give even odds to those two things. But it's something that you'd give even odds to on a bet, even in a deterministic universe.
If I flip a coin and it lands on heads, you'd be a fool to bet otherwise. It doesn't matter if the universe is nondeterministic and you can prove that, given all the knowledge of the universe before the coin was flipped, it would be exactly equally likely to land on heads or tails. You know it landed on heads. It's 100% certain.
Yes, future is uncertain but past is already fixed and certain. So? We are not talking about probabilities of something happening in the past. The topic of the discussion is how temperature (and/or probabilities) are "in the mind" and what does that mean.
The past is certain but the future is not. But the only difference between the two is when you are in relation to them. It's not as if certain time periods are inherently past or future.
An example of temperature being in the mind that's theoretically possible to set up but you'd never manage in practice is Maxwell's demon. If you already know where all of the particles of gas are and how they're bouncing, you could make it so all the fast ones end up in one chamber and all the slow ones end up in the other. Or you can just get all of the molecules into the same chamber. You can do this with an arbitrarily small amount of energy.
I think his "in the mind" is correct in his context, because in the model of entropy he is discussing, temperature_entropy is dependent on entropy, is dependent on your knowledge of the states of the system.
I'll repeat what I said earlier in the context of the discussion of different theories of time.
New physics didn't make old ideas useless. Temperature_kineticenergy is probably more relevant in most situations.
The OP makes his mistake by identifying temperature_entropy with temperature_kineticenergy.
I'm don't see the issue in saying [you don't know what temperature really is] to someone working with the definition [T = average kinetic energy]. One definition of temperature is always true. The other is only true for idealized objects.
Nobody knows what anything really is. We have more or less accurate models.
What do you mean by "true"? They both can be expressed for any object. They are both equal for idealized objects.
Only one of them actually corresponds with temperature for all objects. They are both equal for one subclass of idealized objects, in which case the "average kinetic energy" definition follows from the the entropic definition, not the other way around. All I'm saying is that it's worth emphasizing that one definition is strictly more general than the other.
Average kinetic energy always corresponds to average kinetic energy, and the amount of energy it takes to create a marginal amount of entropy always corresponds to the amount of energy it takes to create a marginal amount of entropy. Each definition corresponds perfectly to itself all of the time, and applies to the other in the case of idealized objects. How is one more general?
Two systems with the same "average kinetic energy" are not necessarily in equilibrium. Sometimes energy flows from a system with lower average kinetic energy to a system with higher average kinetic energy (eg. real gases with different degrees of freedom). Additionally "average kinetic energy" is not applicable at all to some systems, eg. ising magnet.
I just mean as definitions of temperature. There's temperature(from kinetic energy) and temperature(from entropy). Temperature(from entropy) is a fundamental definition of temperature. Temperature(from kinetic energy) only tells you the actual temperature in certain circumstances.
Why is one definition more fundamental than another? Why is only one definition "actual"?
So, effectively there are two different things which go by the same name? Temperature_entropy is one measure (coming from the information-theoretic side) and temperature_kineticenergy is another measure (coming from, um, pre-Hamiltonian mechanics?)..?
That makes some sense, but then I have a question. If you take an ice cube out of the freezer and put it on a kitchen counter, will it melt if there is no one to watch it? In other words, how does the "temperature is in the mind" approach deal with phase transitions?
They look like two different concepts to me.
I don't know. I suppose that would depend on how much that mind knows about phase transitions.
That's difficult to say. If you build a heat pump, you deal with entropy. If you radiate waste heat, you deal with kinetic energy. If you want to know how much waste heat you're going to have, you deal with entropy. If you significantly change the temperature of something with a heat pump, then you have to deal with both for a large variety of temperatures.
Calling them Temperature_kineticenergy and Temperature_entropy is somewhat misleading, since both involve kinetic energy. Temperature_kineticenergy is average kinetic energy, and Temperature_entropy is the change in kinetic energy necessary to cause a marginal increase in entropy.
Also, if you escape your underscores with backslashes, you won't get the italics.