BOOK REVIEW

# The Computational Universe

*Programming the Universe: A Quantum Computer Scientist Takes on
the Cosmos*. Seth Lloyd. xii + 221 pp. Alfred A. Knopf,
2006. $25.95.

In the 1940s, computer pioneer Konrad Zuse began to speculate that
the universe might be nothing but a giant computer continually
executing formal rules to compute its own evolution. He published
the first paper on this radical idea in 1967, and since then it has
provoked an ever-increasing response from popular culture (the film
*The Matrix,* for example, owes a great deal to Zuse's
theories) and hard science alike.

Given this backdrop, Seth Lloyd appears to be exaggerating when he
claims in his informative and entertaining new book that he
"advocates a new paradigm" by postulating the universe to
be a machine that processes information. However, in the book, which
is titled *Programming the Universe,* Lloyd does somewhat
distinguish himself from his predecessors by focusing on the weird
world of quantum computation. He lucidly explains what quantum
computation is all about, how the process of quantum entanglement
seems to involve an instantaneous exchange of information between
locations that can be light-years apart, and why this phenomenon
unfortunately cannot be exploited to transmit information faster
than light. He also describes how quantum computers would be able to
solve certain problems much faster than their traditional counterparts.

The book's central conceit is that the universe can be viewed as a
giant quantum computer made up of connected quantum gates that flip
quantum bits and thereby propagate information and uncertainty in an
"infectious" way. Lloyd uses results by Hans Joachim
Bremermann, Norman Margolus and Lev Levitin to calculate the
processing power of the "ultimate laptop" (one with 1
kilogram of mass and 1 liter of volume): a maximum of
10^{51} operations per second on 10^{32} bits. A
good fan would be needed, though: The massively parallel laptop
would be roughly 100 times hotter than the center of the Sun. Lloyd
also calculates that the visible universe has so far computed about
10^{122} operations on 10^{92} bits. Doesn't sound
like a lot, does it?

Like most quantum physicists, Lloyd believes that there is a source of true randomness, which manifests itself whenever we measure a quantum bit. He does not explicitly mention, however, that this belief has never been experimentally proved. Neither Heisenberg's uncertainty principle nor Bell's inequality (which rests on the very assumption of randomness) excludes the possibility that the universe (including all observers inhabiting it) is in principle computable by a deterministic computer, as first suggested by Zuse.

In fact, Lloyd's belief in true randomness also seems inconsistent with his invocation of Ockham's razor, which favors simple explanations of the universe's history over complex ones. According to both standard and algorithmic information theory, true randomness actually corresponds to maximal information, complexity and description length—the opposite of simplicity.

As long as the (currently somewhat unfashionable) hypothesis of determinism remains unfalsified, many scientists will be dissatisfied by an explanation of our universe's history that requires an enormous amount of information to describe all the random events that have taken place, in addition to the known, compactly describable physical laws. Physicists should keep searching for simple, deterministic, pseudorandom computational rules explaining any type of hitherto-unexplained apparent randomness. Einstein, whose belief that "God does not play dice" has not yet been proved wrong, would probably agree.

The book is least convincing when it comes to the topics of complexity, entropy and algorithmic information. Lloyd compares random events at the quantum level to monkeys typing a random program on the universal computer; this is linked to Ray Solomonoff's basic concept of algorithmic probability theory—namely, that short random programs are more likely than long ones. However, the point of Solomonoff's approach is that some programs can remain short by ceasing to read new input bits. This essential feature seems absent from Lloyd's setup, which demands the permanent creation of new bits corresponding to never-ending programs, thus making each "program" extremely unlikely.

Of all the various complexity measures, Lloyd explicitly favors those that combine the difficulty of doing something with the difficulty of describing it. The obvious choice here would be Leonid A. Levin's Kt-complexity (and its probabilistic derivative, the Speed Prior), which trades off space and computation time in a way that is theoretically optimal in a certain sense. Lloyd, however, does not even mention it. Instead he focuses on measures based on Charles H. Bennett's logical depth, and something called "effective complexity," which reflects a preliminary attempt of physicist Murray Gell-Mann and Lloyd himself to hide the complexity of truly random bits. Generally speaking, the connections between Lloyd's model of quantum processing and algorithmic information theory seem vague.

Some of Lloyd's statements reflect a certain naiveté about
some topics in computer science. For example, he writes that
"According to the Church-Turing hypothesis, every possible
mathematical structure is represented in some component of the
superposition" of all possible computations, despite the fact
that much of mathematics deals with incomputable objects. Other
incorrect statements include this one: "In fact, every
universal computer can be shown not only to simulate every other
universal computer, but to do so *efficiently."*

Lloyd's historical notes on computation and bits refer to Charles Babbage's analytic engine and John Napier's logarithmic bones but fail to mention Gottfried Wilhelm Leibniz, the inventor of the bit (1700), and Wilhelm Schickard, constructor of the very first (non-program-controlled) computer, in 1623. For some reason Lloyd also seems to give equal credit to Zuse and Ed Fredkin as creators of the "universe as a computer" idea, although Fredkin got into this business long after Zuse. On the other hand, Lloyd did enrich my understanding by pointing out that the "many worlds" theory (usually attributed to the physicist Hugh Everett) can be traced back to poet and novelist Jorge Luis Borges.

Lloyd spices his story with interesting and sometimes touching personal tales of his career at the border between computer science and physics. Despite my few quibbles, I recommend this well-written book without hesitation to anybody interested in an overview of basic ideas in the field. I intend to buy a few copies as presents for my friends.

**IN THIS SECTION**

Community Guidelines: Disqus Comments

**BROWSE BY**

- Nanoview

- Reviewer

- Topic

- Issue

# Connect With Us:

# Subscribe to Free eNewsletters!

*American Scientist Update: Meat-Eating Early Humans, Infant Vision, Spring's Budburst, and More Science Stories**Scientists' Nightstand: Holiday Special!*News of book reviews published in

An early peek at each new issue, with descriptions of feature articles, columns, and more. Issues contain links to everything in the latest issue's table of contents.

*American Scientist*and around the web, as well as other noteworthy happenings in the world of science books.

To sign up for automatic emails of the

*American Scientist Update*and

*Scientists' Nightstand*issues, create an online profile, then sign up in the My AmSci area.

# RSS Feed Subscription

Receive notification when new content is posted from the entire website, or choose from the customized feeds available.

# Read Past Issues on JSTOR

JSTOR, the online academic archive, contains complete back issues of *American Scientist* from 1913 (known then as the *Sigma Xi Quarterly*) through 2005.

The table of contents for each issue is freely available to all users; those with institutional access can read each complete issue.

View the full collection here.

EMAIL TO A FRIEND :

**Of Possible Interest**

**Book Excerpt**: Cypherpunks Write Code

**Book Review**: Names, Simplified

**Book Review**: Don't Try This at Home