Skip to main content

Has quantum computing been cracked?

In recent days there has been a surge in interest in quantum computing - computers that use quantum particles as the equivalent of bits. Out of the blue, I've received several invitations to talk to people about quantum computing as a result of my my book, imaginatively named Quantum Computing, which provides an introduction to the field. I suspect this upsurge is because of the recent announcement that the BBC dramatically headlined Quantum breakthrough could revolutionise computing

This is a topic that has suffered from considerable hype in the past - so is this breakthrough (which there certainly has been) transformative or an incremental step towards what is still a fairly distant proposition?

The reason quantum computers are of huge interest is that for certain applications they can, in principle, carry out calculations that would take conventional computers the lifetime of the universe to churn through. The reason that they can do this is that instead of using bits that can store values of 0 and 1, the quantum computer uses qubits - each a quantum particle which can be in a superposition of states - partly 0 and partly 1 simultaneously, with the 'partly' effectively capable of representing an infinitely long real value. The way that qubits link together means that what would usually require sequential processes in a conventional computer can be undertaken simultaneously.

However, there also plenty of problems with making quantum computers work. You need to be able to isolate quantum particles from their environment, or the states of the qubits will be lost, while still being able to interact with them. This is not trivial and as yet it has limited quantum computers to orders of magnitude around 100 qubits. You also need to undertake error correction, because the process is inherently prone to errors, which means it takes considerably more qubits to undertake a calculation that might otherwise be thought. What's more, you need to have both a suitable algorithm, specifically devised for a quantum computer, and the ability to get information in and out of the computer, when the typical answer may well just be 0 or 1.

It's important to emphasise that quantum computers are not desktop devices - they may well always require a specially controlled environment, working as shared cloud devices - and they are not general purpose computers, with relatively limited numbers of potentially very powerful algorithms. The first two examples  produced were an algorithm that effectively makes it easier to crack the encryption used for internet payments (a trifle worrying), and (the reason Google, for example, is very interested) a search algorithm that makes it possible to find something with the square root of the number of searches required by a conventional computer. To emphasise how much the development of this hardware is a slow process, these algorithms were both developed in the mid-1990s, long before anything was available to run them on.

The breakthrough that is making the news involves one class of quantum computers - those where the qubits are based on ions (atoms that have gained or lost electrons to become electrically charged). Other quantum computers use photons, for example, but ions have the advantage of being relatively easy to keep in place due to their electrical charge. A chip to confine and interact with ions requires a lot more space that dealing with the equivalent number of conventional bits. A standard-sizes chip can only handle around 100 qubits, where an effective quantum computer might require a few millions (still vastly smaller than the billions of bits in a conventional computer processor). The breakthrough involves being able to transfer ions from one chip to another with a very low loss rate and without measurably impacting the 'phase coherence' of the qubit - in simple terms, the qubit keeps the values its holding.

This is an impressive piece of work. It makes it possible in principle to have a quantum computer with many chips that interact with each other, enabling it to support the kind of number of qubits that would make it a truly effective resource. However, it's worth emphasising that there are still plenty of other issues to be dealt with, and that while this is an effective demonstration, it's still a way from being applicable on any scale. Realistically it could be another 5 to 10 years before there is a real product where large scale, useful quantum algorithms can be deployed. An important step, then, but definitely incremental rather than a revolution.

If you'd like to read more about the technology, the paper is here and is freely downloadable. (Surely it's time the BBC started providing links to papers?)


Comments

Popular posts from this blog

The Decline and Fall of the Human Empire - Henry Gee ****

In his last book, Henry Gee impressed with his A (Very) Short History of Life on Earth - this time he zooms in on one very specific aspect of life on Earth - humans - and gives us not just a history, but a prediction of the future - our extinction. The book starts with an entertaining prologue, to an extent bemoaning our obsession with dinosaurs, a story that leads, inexorably towards extinction. This is a fate, Gee points out, that will occur for every species, including our own. We then cover three potential stages of the rise and fall of humanity (the book's title is purposely modelled on Gibbon) - Rise, Fall and Escape. Gee's speciality is palaeontology and in the first section he takes us back to explore as much as we can know from the extremely patchy fossil record of the origins of the human family, the genus Homo and the eventual dominance of Homo sapiens , pushing out any remaining members of other closely related species. As we move onto the Fall section, Gee gives ...

Pagans (SF) - James Alistair Henry *****

There's a fascinating sub-genre of science fiction known as alternate history. The idea is that at some point in the past, history diverged from reality, resulting in a different present. Perhaps the most acclaimed of these books is Kingsley Amis's The Alteration , set in a modern England where there had not been a reformation - but James Alistair Henry arguably does even better by giving us a present where Britain is a third world country, still divided between Celts in the west and Saxons in the East. Neither the Normans nor Christianity have any significant impact. In itself this is a clever idea, but what makes it absolutely excellent is mixing in a police procedural murder mystery, where the investigation is being undertaken by a Celtic DI, Drustan, who has to work in London alongside Aedith, a Saxon reeve of equivalent rank, who also happens to be daughter of the Earl of Mercia. While you could argue about a few historical aspects, it's effectively done and has a plot...

Amazing Worlds of Science Fiction and Science Fact: Keith Cooper ****

There's something appealing (for a reader like me) about a book that brings together science fiction and science fact. I had assumed that the 'Amazing Worlds' part of the title suggested a general overview of the interaction between the two, but Keith Cooper is being literal. This is an examination of exoplanets (planets that orbit a different star to the Sun) as pictured in science fiction and in our best current science, bearing in mind this is a field that is still in the early phases of development. It becomes obvious early on that Cooper, who is a science journalist in his day job, knows his stuff on the fiction side as well as the current science. Of course he brings in the well-known TV and movie tropes (we get a huge amount on Star Trek ), not to mention the likes of Dune, but his coverage of written science fiction goes into much wider picture. He also has consulted some well-known contemporary SF writers such as Alastair Reynolds and Paul McAuley, not just scient...