![]() |
|
![]() |
| Euler.
JVM was one of the smartest ever, but Euler was there centuries before and shows up in so many places. If I had a Time Machine I'd love to get those two together for a stiff drink and a banter. |
![]() |
| Cross entropy lets you compare two probability distributions. One way you can apply it is to let the distribution p represent "reality" (from which you can draw many samples, but whose numerical value you might not know) and to let q represent "beliefs" (whose numerical value is given by a model). Then by finding q to minimize cross-entropy H[p, q] you can move q closer to reality.
You can apply it other ways. There are lots of interpretations and uses for these concepts. Here's a cool blog post if you want to find out more: https://blog.alexalemi.com/kl-is-all-you-need.html |
![]() |
| Entropy in physics is usually the Shannon entropy of the probability distribution over system microstates given known temperature and pressure. If the system is in equilibrium then this is objective. |
![]() |
| Sounds like a non-sequitur to me; what are you implying about the Maxwell's demon thought experiment vs the comparison between Shannon and stat-mech entropy? |
![]() |
| To shorten this for you with my own (identical) understanding: "entropy is just the name for the bits you don't have".
Entropy + Information = Total bits in a complete description. |
![]() |
| Trivial example: if you know the seed of a pseudo-random number generator, a sequence generated by it has very low entropy.
But if you don't know the seed, the entropy is very high. |
![]() |
| Can you explain this in more detail?
Entropy is calculated as a function of a probability distribution over possible messages or symbols. The sender might have a distribution P over possible symbols, and the receiver might have another distribution Q over possible symbols. Then the "true" distribution over possible symbols might be another distribution yet, call it R. The mismatch between these is what leads to various inefficiencies in coding, decoding, etc [1]. But both P and Q are beliefs about R -- that is, they are properties of observers. [1] https://en.wikipedia.org/wiki/Kullback–Leibler_divergence#Co... |
![]() |
| The second law of thermodynamics is time-asymmetric, but the fundamental physical laws are time-symmetric, so from them you can only predict that the entropy of B should be bigger than the entropy of A irrespective of whether B is in the future or the past of A. You need the additional assumption (Past Hypothesis) that the universe started in a low entropy state in order to get the second law of thermodynamics.
> If our goal is to predict the future, it suffices to choose a distribution that is uniform in the Liouville measure given to us by classical mechanics (or its quantum analogue). If we want to reconstruct the past, in contrast, we need to conditionalize over trajectories that also started in a low-entropy past state — that the “Past Hypothesis” that is required to get stat mech off the ground in a world governed by time-symmetric fundamental laws. https://www.preposterousuniverse.com/blog/2013/07/09/cosmolo... |
![]() |
| Hey did you want to say relative entropy ~ rate function ~ KL divergence. Might be more familiar to ML enthusiasts here, get them to be curious about Sanov or large deviations. |
![]() |
| Ah JCB, how I love your writing, you are always so very generous.
Your This Week's Finds were a hugely enjoyable part of my undergraduate education and beyond. Thank you again. |
![]() |
| Thanks for the reference will take some time before I see the whole video. Can you tell me what those quantum fluctuations are in short? Are they part of some physical law? |
![]() |
| Symmetry breaking is the general phenomenon that underlies most of that.
The classic example is this: Imagine you have a perfectly symmetrical sombrero[1], and there's a ball balanced on top of the middle of the hat. There's no preferred direction it should fall in, but it's _unstable_. Any perturbation will make it roll down hill and come to rest in a stable configuration on the brim of the hat. The symmetry of the original configuration is now broken, but it's stable. 1: https://m.media-amazon.com/images/I/61M0LFKjI9L.__AC_SX300_S... |
![]() |
| After years of thought I dare to say the 2nd TL is a tautology. Entropy is increasing means every system tends to higher probability means the most probable is the most probable. |
![]() |
| The way I understand it is with an analogy to probability. To me, events are to microscopic states like random variable is to entropy. |
![]() |
| Correct! And it took me just one paragraph, not the 18 pages of meandering (and I think confusing) text that it takes the author of the pdf to introduce the same idea. |
![]() |
| You didn’t introduce any idea. You said it’s “just a number” and wrote down a formula without any explanation or justification.
I concede that it was much shorter though. Well done! |
![]() |
| This seems like a great resource for referencing the various definitions. I've tried my hand at developing an intuitive understanding: https://spacechimplives.substack.com/p/observers-and-entropy. TLDR - it's an artifact of the model we're using. In the thermodynamic definition, the energy accounted for in the terms of our model is information. The energy that's not is entropic energy. Hence why it's not "useable" energy, and the process isn't reversible.
|
![]() |
| There's a good case to be made that the information-theoretic definition of entropy is the most fundamental one, and the version that shows up in physics is just that concept as applied to physics. |
![]() |
| Well it's part of math, which physics is already based on.
Whereas metaphysics is, imo, "stuff that's made up and doesn't matter". Probably not the most standard take. |
"My greatest concern was what to call it. I thought of calling it 'information,' but the word was overly used, so I decided to call it 'uncertainty.' When I discussed it with John von Neumann, he had a better idea. Von Neumann told me, 'You should call it entropy, for two reasons. In the first place your uncertainty function has been used in statistical mechanics under that name, so it already has a name. In the second place, and more important, no one really knows what entropy really is, so in a debate you will always have the advantage.'"
See the answers to this MathOverflow SE question (https://mathoverflow.net/questions/403036/john-von-neumanns-...) for references on the discussion whether Shannon's entropy is the same as the one from thermodynamics.