It’s true that entropy is one of the more difficult concepts in physics, at least as regards its precise scientific meaning. At an intuitive level, however, even people who have never heard the word are probably aware of it. If you’ve ever had a sneaking suspicion that energy might not actually be conserved in the way science teachers say it is – because in your experience things are always running down and wearing out – then that’s a consequence of entropy. In simple terms, it can be thought of as a measure of how much ‘useless’ energy there is in a closed system, in the sense of energy that can’t be exploited to any practical purpose. And entropy, unlike energy, isn’t conserved – it can increase, but it can never decrease. If conservation of energy is the good news, then entropy is the bad news. As Binney points out more than once in the book, energy is free – it’s in abundant supply all around us. What costs money, and causes so many headaches for politicians and the public alike, is low-entropy energy that can actually be used for something.
Rewinding to the end of the first paragraph, you may spot something a little odd. If James Binney knows there are so few people that have even heard of entropy, why has he used it as a single-word title of this book? The answer is quite simple: it’s part of OUP’s long-running ‘Very Short Introduction’ series, which operates within well-established constraints – one of these being that the title is always a straight description of the subject matter. As to the style of the books themselves, they are indeed short – this one is only a little over a hundred small pages – and they are ‘introductions’, in the sense of assuming the reader starts with zero knowledge of the subject. As a rule, though, they’re aimed at serious readers who need a quick but thorough-going crash course – whether that’s to pass an exam or for a work-related project – rather than general readers looking for a few entertaining and memorable highlights. In the context of a science subject like this one, the result isn’t really what you would understand by the term ‘popular science’, and was never meant to be.
As regards this specific book, casual readers will find it bombards them with far more technical detail than they wanted to know about the mathematical definitions of entropy, its relationship to other areas of physics, and the way that scientists and engineers use it in their everyday work. Even so, there are some fascinating insights, particularly on the historical development of the subject – with various eminent scientists spending several decades of the nineteenth century trying to unscramble their understanding of energy and entropy – as well as its relationship to cutting-edge areas of modern physics such as quantum entanglement and black holes. There are also a few interesting examples – though not as many as I would have liked – of what might be called ‘everyday’ applications of entropy, for example in heat pumps and image processing.
The very title of this book, and the fact that it’s part of this particular OUP series, means that in practice anyone who buys it will have a pretty good idea what they’re letting themselves in for. It’s these readers that my 4-star rating is aimed at. On the other hand, if it happened to come into the hands of a more general popular science reader, they’re likely to find far more technical detail than they are used to. If they’re happy to skip over this, however, they’ll still find plenty of interesting things that aren’t widely covered elsewhere.
Review by Andrew May - See all Brian's online articles or subscribe to a weekly email free here



Comments
Post a Comment