Skip to main content

Everything is Predictable - Tom Chivers *****

There's a stereotype of computer users: Mac users are creative and cool, while PC users are businesslike and unimaginative. Less well-known is that the world of statistics has an equivalent division. Bayesians are the Mac users of the stats world, where frequentists are the PC people. This book sets out to show why Bayesians are not just cool, but also mostly right.

Tom Chivers does an excellent job of giving us some historical background, then dives into two key aspects of the use of statistics. These are in science, where the standard approach is frequentist and Bayes only creeps into a few specific applications, such as the accuracy of medical tests, and in decision theory where Bayes is dominant.

If this all sounds very dry and unexciting, it's quite the reverse. I admit, I love probability and statistics, and I am something of a closet Bayesian*), but Chivers' light and entertaining style means that what could have been the mathematical equivalent of debating angels on the heads of a pin becomes both enthralling and relatively easy to understand. You may have to re-read a few sentences, because there is a bit of a head-scrambling concept at the heart of the debate - but it's well worth it.

A trivial way of representing the difference between Bayesian and frequentist statistics is how you respond to the question 'What's the chance of the result being a head?' when looking at a coin that has already been tossed, but that you haven't seen. Bayesian statistics takes into account what you already know. As you don't know what the outcome is, you can only realistically say it's 50:50, or 0.5 in the usual mathematical representation. By contrast, frequentist statistics says that as the coin has been tossed, it is definitely heads or tails with probability 1... but we can't say which. This seems perhaps unimportant - but the distinction becomes crucial when considering the outcome of scientific studies.

Thankfully, Chivers goes into in significant detail the problem that arises because in most scientific use of (frequentist) probability, what the results show is not what we actually want to know. In the social sciences, a marker for a result being 'significant' is a p-value of less that 0.05. This means that if the null hypothesis is true (the effect you are considering doesn't exist), then you would only get this result 1 in 20 times or less. But what we really want to know is not the chance of this result if the hypothesis is true, but rather what's the chance that the hypothesis is true - and that's a totally different thing.

Chivers gives the example of 'it's the difference between "There's only a 1 in 8 billion chance that a given human is the Pope" and "There's only a 1 in 8 billion chance that the Pope is human"'. At risk of repetition because it's so important, frequentist statistics, as used by most scientists, tells us the chance of getting the result if the hypothesis is true; Bayesian statistics works out what the chance is of the hypothesis being true - which most would say is what we really want to know. In fact, as Chivers points out, most scientists don't even know that they aren't showing the chance of the hypothesis being true - and this even true of many textbooks for scientists on how to use statistics.

At this point, most normal humans would say 'Why don't those stupid scientists use Bayes?' But there is a catch. To be able to find how likely the hypothesis is, we need a 'prior probability' - a starting point which Bayes' theorem then modifies using the evidence we have. This feels subjective, and for the first attempt at a study it certainly can be. But, as Chivers points out, in many scientific studies there is existing evidence to provide that starting point - the frequentist approach throws away this useful knowledge.

Is the book perfect? Well, I suspect as a goodish Bayesian I can never say something is perfect. I found it hard to engage with an overlong chapter called 'the Bayesian brain' that is not about using Bayes, but rather trying to show that our brains take this approach, which all felt a bit too hypothetical for me. And Chivers repeats the oft-seen attack on poor old Fred Hoyle, taking his comment about evolution and 'a whirlwind passing through a junkyard creating a Boeing 747' in a way that oversimplifies Hoyle's original meaning. But these are trivial concerns.

I can't remember when I last enjoyed a popular maths book so much. It's a delight.

* Not entirely a closet Bayesian - my book Dice World includes an experiment using Bayesian statistics to work out what kind of dog I have, given a mug that's on my desk.

Hardback:   
Kindle 
Using these links earns us commission at no cost to you
Review by Brian Clegg - See all Brian's online articles or subscribe to a weekly email free here

Comments

Popular posts from this blog

It's On You - Nick Chater and George Loewenstein *****

Going on the cover you might think this was a political polemic - and admittedly there's an element of that - but the reason it's so good is quite different. It shows how behavioural economics and social psychology have led us astray by putting the focus way too much on individuals. A particular target is the concept of nudges which (as described in Brainjacking ) have been hugely over-rated. But overall the key problem ties to another psychological concept: framing. Huge kudos to both Nick Chater and George Loewenstein - a behavioural scientist and an economics and psychology professor - for having the guts to take on the flaws in their own earlier work and that of colleagues, because they make clear just how limited and potentially dangerous is the belief that individuals changing their behaviour can solve large-scale problems. The main thesis of the book is that there are two ways to approach the major problems we face - an 'i-frame' where we focus on the individual ...

Introducing Artificial Intelligence – Henry Brighton & Howard Selina ****

It is almost impossible to rate these relentlessly hip books – they are pure marmite*. The huge  Introducing  … series (a vast range of books covering everything from Quantum Theory to Islam), previously known as …  for Beginners , puts across the message in a style that owes as much to Terry Gilliam and pop art as it does to popular science. Pretty well every page features large graphics with speech bubbles that are supposed to emphasise the point. Funnily,  Introducing Artificial Intelligence  is both a good and bad example of the series. Let’s get the bad bits out of the way first. The illustrators of these books are very variable, and I didn’t particularly like the pictures here. They did add something – the illustrations in these books always have a lot of information content, rather than being window dressing – but they seemed more detached from the text and rather lacking in the oomph the best versions have. The other real problem is that...

The Laws of Thought - Tom Griffiths *****

In giving us a history of attempts to explain our thinking abilities, Tom Griffiths demonstrates an excellent ability to pitch information just right for the informed general reader.  We begin with Aristotelian logic and the way Boole and others transformed it into a kind of arithmetic before a first introduction of computing and theories of language. Griffiths covers a surprising amount of ground - we don't just get, for instance, the obvious figures of Turing, von Neumann and Shannon, but the interaction between the computing pioneers and those concerned with trying to understand the way we think - for example in the work of Jerome Bruner, of whom I confess I'd never heard.  This would prove to be the case with a whole host of people who have made interesting contributions to the understanding of human thought processes. Sometimes their theories were contradictory - this isn't an easy field to successfully observe - but always they were interesting. But for me, at least, ...