Skip to main content

Everything is Predictable - Tom Chivers *****

There's a stereotype of computer users: Mac users are creative and cool, while PC users are businesslike and unimaginative. Less well-known is that the world of statistics has an equivalent division. Bayesians are the Mac users of the stats world, where frequentists are the PC people. This book sets out to show why Bayesians are not just cool, but also mostly right.

Tom Chivers does an excellent job of giving us some historical background, then dives into two key aspects of the use of statistics. These are in science, where the standard approach is frequentist and Bayes only creeps into a few specific applications, such as the accuracy of medical tests, and in decision theory where Bayes is dominant.

If this all sounds very dry and unexciting, it's quite the reverse. I admit, I love probability and statistics, and I am something of a closet Bayesian*), but Chivers' light and entertaining style means that what could have been the mathematical equivalent of debating angels on the heads of a pin becomes both enthralling and relatively easy to understand. You may have to re-read a few sentences, because there is a bit of a head-scrambling concept at the heart of the debate - but it's well worth it.

A trivial way of representing the difference between Bayesian and frequentist statistics is how you respond to the question 'What's the chance of the result being a head?' when looking at a coin that has already been tossed, but that you haven't seen. Bayesian statistics takes into account what you already know. As you don't know what the outcome is, you can only realistically say it's 50:50, or 0.5 in the usual mathematical representation. By contrast, frequentist statistics says that as the coin has been tossed, it is definitely heads or tails with probability 1... but we can't say which. This seems perhaps unimportant - but the distinction becomes crucial when considering the outcome of scientific studies.

Thankfully, Chivers goes into in significant detail the problem that arises because in most scientific use of (frequentist) probability, what the results show is not what we actually want to know. In the social sciences, a marker for a result being 'significant' is a p-value of less that 0.05. This means that if the null hypothesis is true (the effect you are considering doesn't exist), then you would only get this result 1 in 20 times or less. But what we really want to know is not the chance of this result if the hypothesis is true, but rather what's the chance that the hypothesis is true - and that's a totally different thing.

Chivers gives the example of 'it's the difference between "There's only a 1 in 8 billion chance that a given human is the Pope" and "There's only a 1 in 8 billion chance that the Pope is human"'. At risk of repetition because it's so important, frequentist statistics, as used by most scientists, tells us the chance of getting the result if the hypothesis is true; Bayesian statistics works out what the chance is of the hypothesis being true - which most would say is what we really want to know. In fact, as Chivers points out, most scientists don't even know that they aren't showing the chance of the hypothesis being true - and this even true of many textbooks for scientists on how to use statistics.

At this point, most normal humans would say 'Why don't those stupid scientists use Bayes?' But there is a catch. To be able to find how likely the hypothesis is, we need a 'prior probability' - a starting point which Bayes' theorem then modifies using the evidence we have. This feels subjective, and for the first attempt at a study it certainly can be. But, as Chivers points out, in many scientific studies there is existing evidence to provide that starting point - the frequentist approach throws away this useful knowledge.

Is the book perfect? Well, I suspect as a goodish Bayesian I can never say something is perfect. I found it hard to engage with an overlong chapter called 'the Bayesian brain' that is not about using Bayes, but rather trying to show that our brains take this approach, which all felt a bit too hypothetical for me. And Chivers repeats the oft-seen attack on poor old Fred Hoyle, taking his comment about evolution and 'a whirlwind passing through a junkyard creating a Boeing 747' in a way that oversimplifies Hoyle's original meaning. But these are trivial concerns.

I can't remember when I last enjoyed a popular maths book so much. It's a delight.

* Not entirely a closet Bayesian - my book Dice World includes an experiment using Bayesian statistics to work out what kind of dog I have, given a mug that's on my desk.

Hardback:   
Kindle 
Using these links earns us commission at no cost to you
Review by Brian Clegg - See all Brian's online articles or subscribe to a weekly email free here

Comments

Popular posts from this blog

The Art of Uncertainty - David Spiegelhalter *****

There's something odd about this chunky book on probability - the title doesn't mention the P word at all. This is because David Spiegelhalter (Professor Sir David to give him his full title) has what some mathematicians would consider a controversial viewpoint. As he puts it 'all probabilities are judgements expressing personal uncertainty.' He strongly (and convincingly) argues that while the mathematical approach to probability is about concrete, factual values, outside of the 'natural' probabilities behind quantum effects, almost all real world probability is a subjective experience, better described by more subjective terms like uncertainty, chance and luck. A classic way to distinguish between those taking the frequentist approach to probability and the Bayesian approach is their attitude to what the probability is of a fair coin coming up heads or tails after the coin has been tossed but before we have looked at it. The frequentist would say it's def

Math for English Majors - Ben Orlin *****

Ben Orlin makes the interesting observation that the majority of people give up on understanding maths at some point, from fractions or algebra all the way through to tensors. At that stage they either give up entirely or operate the maths mechanically without understanding what they are doing. In this light-hearted take, Orlin does a great job of taking on mathematical processes a step at a time, in part making parallels with the structure of language. Many popular maths books shy away from the actual mathematical representations, going instead for verbal approximations. Orlin doesn't do this, but makes use of those linguistic similes and different ways of looking at the processes involved to help understanding. He also includes self-admittedly awful (but entertaining) drawings and stories from his experience as a long-time maths teacher. To make those parallels, Orlin refers to numbers as nouns, operations as verbs (though he points out that there are some flaws in this simile) a