Skip to main content

Nature’s Nanotech #4 – The importance of being wet – Brian Clegg


Fourth in our Nature’s Nanotech series
The image that almost always springs to mind when nanotechnology is mention is Drexler’s tiny army of assemblers and the threat of being overwhelmed by grey goo. But what many forget is that there is a fundamental problem in physics facing anyone building invisibly small robots (nanobots) – something that was spotted by the man who first came up with the concept of working on the nanoscale.
That man was Richard Feynman. His name may not be as well known outside physics circles as, say, Stephen Hawking, but ask a physicist to add a third to a triumvirate of heroes with Newton and Einstein and most would immediately choose Feynman. It didn’t hurt that Richard Feynman was a bongo-playing charmer whose lectures delighted even those who couldn’t understand the science, helped by an unexpected Bronx accent – imagine Tony Curtis lecturing on quantum theory.
Feynman became best known to the media for his dramatic contribution to the Challenger inquiry, when in front of the cameras he plunged an O-ring into iced water to show how it lost its elasticity. But on an evening in December 1959 he gave a lecture that laid the foundation for all future ideas of nanobots. His talk at the annual meeting of the American Physical Society was titled There’s Plenty of Room at the Bottom, and his subject was manipulating and controlling things on a small scale.
Feynman pointed out that people were amazed by a device that could write the Lord’s Prayer on the head of a pin. But ‘Why cannot we write the entire 24 volumes of the Encyclopedia Britannica on the head of a pin?’ As he pointed out, the dots that make up a printed image, if reduced to a scale that took the area of paper in the encyclopedia down to pinhead size, would still contain 1,000 atoms each – plenty of material to make a pixel. And it could be read with technology they had already.
Feynman went on to describe how it would be possible to write at this scale, but also took in the idea that the monster computers of his day would have to become smaller and smaller to cram in the extra circuits required for sophisticated computation. Then he described how engineering could be undertaken on the nanoscale, and to do so, he let his imagination run a little wild.
What Feynman envisaged was making use of the servo ‘hands’ found in nuclear plants to act remotely, but instead of making the hands the same size as the original human hands, building them on a quarter scale. He would also construct quarter size lathes to produce scaled down parts for new devices. These quarter scale tools would be used to produce sixteenth scale hands and lathes, which themselves would produce sixty-fourth scale items… and so on, until reaching the nanoscale.
The second component of Feynman’s vision was a corresponding multiplication of quantity, as you would need billions of nanobots to do anything practical. So he would not make one set of quarter scale hands, but ten. Each of those would produce 10 sixteenth scale devices, so there would be 100 of them – and so on. Feynman points out there would not be a problem of space or materials, because one billion 1/4000 scale lathes would only take up two percent of the space and materials of a conventional lathe.
When he discussed running nanoscale machines, Feynman even considered the effect on lubrication. The mechanical devices we are familiar with need oil to prevent them ceasing up. As he pointed out, the effective viscosity of oil gets higher and higher in proportion as you go down in scale. It stops being a lubricant and starts being like attempting to operate in a bowl of tar. But, he argues, you may well not need lubricants, as the bearings won’t run hot because the heat would escape very rapidly from such a small device.
So far, so good, but what is the problem Feynman mentions? He points out that ‘As we go down in size there are a number of interesting problems that arise. All things do not simply scale down in proportion.’ Specifically, as things get smaller they begin to stick together. If you unscrewed a nanonut from a nanobolt it wouldn’t fall off – the Van der Waals force we met on the gecko’s foot is stronger than the force of gravity on this scale. Small things stick together in a big way.
Feynman is aware there would be problems. ‘It would be like those old movies of a man with his hands full of molasses, trying to get rid of a glass of water.’ But he does effectively dismiss the problems. In reality, the nano-engineer doesn’t just have Van der Waals forces to deal with. Mechanical engineering generally involves flat surfaces briefly coming together to transfer force from one to the other, as when the teeth of a pair of gears mesh. But down at the nanoscale a new, almost magical, force springs into life – the Casimir effect.
If two plates get very close, they are attracted towards each other. This has nothing to do with electromagnetism, like the Van der Waals force, but is the result of a weird aspect of quantum theory. All the time, throughout all of space, quantum particles briefly spring into existence, then annihilate each other. An apparently empty vacuum is, in fact, a seething mass of particles that exist for such a short space of time that we don’t notice them.
However, one circumstance when these particles do come to the fore is when there are two sheets of material very close to each other. If the space separating the sheets is close enough, far fewer of these ‘virtual’ particles can appear between them than outside them. The result is a real pressure that pushes the plates together. Tiny parallel surfaces slam together under this pressure.
The result of these effects is that even though toy nanoscale gears have been constructed from atoms, a real nanotechnology machine – a nanobot – would simply not work using conventional engineering. Instead the makers of nanobots need to look to nature. Because the natural world has plenty of nanoscale machines, moving around, interacting and working. What’s the big difference? Biological machines are wet and soft.
By this I don’t mean they use water as a lubricant rather than oil, but rather they are not usually a device made up of a series of interlocking mechanical components like our machines but rather use a totally different approach to mechanisms and interaction that results in a ‘wet’, soft environment lacking flat surfaces and the opportunities for small scale stickiness to get in the way of their workings.
If we are to build nanomachines, our engineers need to think in a totally different way. We need to dismiss Feynman’s picture of miniature lathes, nuts, bolts and gears. Instead our model has to be the natural world and the mechanisms that evolution has generated to make our, admittedly inefficient, but still functioning nanoscale technology work and thrive. The challenge is huge – but so is the potential.
In the next article in this series we will look at the lessons we can learn from a specific example of nature’s ability to manufacture technology on the nanoscale – the remarkable virus.

Comments

Popular posts from this blog

Philip Ball - How Life Works Interview

Philip Ball is one of the most versatile science writers operating today, covering topics from colour and music to modern myths and the new biology. He is also a broadcaster, and was an editor at Nature for more than twenty years. He writes regularly in the scientific and popular media and has written many books on the interactions of the sciences, the arts, and wider culture, including Bright Earth: The Invention of Colour, The Music Instinct, and Curiosity: How Science Became Interested in Everything. His book Critical Mass won the 2005 Aventis Prize for Science Books. Ball is also a presenter of Science Stories, the BBC Radio 4 series on the history of science. He trained as a chemist at the University of Oxford and as a physicist at the University of Bristol. He is also the author of The Modern Myths. He lives in London. His latest title is How Life Works . Your book is about the ’new biology’ - how new is ’new’? Great question – because there might be some dispute about that! Many

Stephen Hawking: Genius at Work - Roger Highfield ****

It is easy to suspect that a biographical book from highly-illustrated publisher Dorling Kindersley would be mostly high level fluff, so I was pleasantly surprised at the depth Roger Highfield has worked into this large-format title. Yes, we get some of the ephemera so beloved of such books, such as a whole page dedicated to Hawking's coxing blazer - but there is plenty on Hawking's scientific life and particularly on his many scientific ideas. I've read a couple of biographies of Hawking, but I still came across aspects of his lesser fields here that I didn't remember, as well as the inevitable topics, ranging from Hawking radiation to his attempts to quell the out-of-control nature of the possible string theory universes. We also get plenty of coverage of what could be classified as Hawking the celebrity, whether it be a photograph with the Obamas in the White House, his appearances on Star Trek TNG and The Big Bang Theory or representations of him in the Simpsons. Ha

The Blind Spot - Adam Frank, Marcelo Gleiser and Evan Thompson ****

This is a curate's egg - sections are gripping, others rather dull. Overall the writing could be better... but the central message is fascinating and the book gets four stars despite everything because of this. That central message is that, as the subtitle says, science can't ignore human experience. This is not a cry for 'my truth'. The concept comes from scientists and philosophers of science. Instead it refers to the way that it is very easy to make a handful of mistakes about what we are doing with science, as a result of which most people (including many scientists) totally misunderstand the process and the implications. At the heart of this is confusing mathematical models with reality. It's all too easy when a mathematical model matches observation well to think of that model and its related concepts as factual. What the authors describe as 'the blind spot' is a combination of a number of such errors. These include what the authors call 'the bifur