At the time of writing, two popular science topics have become so ubiquitous that it's hard to find anything new to say about them - neuroscience and artificial intelligence. Almost all the (many) AI books I've read have either been paeans to its wonders or dire warnings of how AI will take over the world or make opaque and biassed decisions that destroy lives. What is really refreshing about Ben Shneiderman's book is that it does neither of these - instead it provides an approach to benefit from AI without suffering the negative consequences. That's why it's an important piece of work.
To do this, Shneiderman takes us right back to the philosophical contrast between rationalism and empiricism. Rationalism, we discover, is driven by rules, logic and well-defined boundaries. Empiricists drive their understanding from observation of the real world where things are more fuzzy. Shneiderman then expands this distinction to that between science and innovation. Here, science is seen on focussing on the the essence of what is happening, while innovation is driven by applications.
When we get to AI, Shneiderman argues that many AI researchers take the science approach - they want to understand how people think and to reproduce human-like intelligence in computers and human-like robots. The empirical, innovation-driven AI researchers, meanwhile, focus on ways that AI can not duplicate and supplant human abilities, but support them. It's the difference between providing a human replacement and an AI-driven super tool that enables the human to work far better. Although Shneiderman makes an effort to portray both sides fairly, there is no doubt that he comes down strongly on the empirical, innovation-driven side - human-centred AI. It is exploring this distinction that makes the book important. Shneiderman argues convincingly that we need to move from AI taking decisions and actions, replacing humans, to human-centred AI that augments human abilities.
Quite a lot of this is driven by the importance of the human-computer interface. Science-driven AI tends to have poor or non-existent user interface, with the AI's processes opaque and impossible to control, where innovation driven-AI puts a lot of importance on having meaningful controls and interface. It's frustrating, then, that someone so strong on good user interface produces a book that has such a bad one - instead of the narrative structure of good writing, Human-Centred AI has the dire, rigid structure of a business book or textbook. We get sections with an opening summary, then an introductory chapter that tells you what the section is going to tell you, then a bit of useful content, before a closing chapter that summarises the section. There is so much repetition of the basic points that it becomes really irritating. The interface of cameras on smartphones, for example, are used as exemplars almost word for word many times over.
The useful content could be covered in a couple of magazine articles - yet when you hit the good stuff it is really good stuff. This is by no means the best way of putting the information across - nevertheless, by dint of this valuable message, it is one of the most important AI books of the last few years.
Hardback: |
Bookshop.org |
Comments
Post a Comment