“It doesn’t matter whether something consists of equations, words, images or sounds—you can encode any of that in strings of zeroes and ones,” or bits, says Scott Aaronson, associate professor of electrical engineering and computer science at MIT. Your computer is doing it right now, using tiny magnets, capacitors, and transistors to store billions or trillions of binary digits. “These might have been hard concepts for people to understand a century ago, but because of the computer revolution, we deal with these concepts all the time,” says Aaronson. In an age when a USB drive dangles from every keychain and an iPhone strains the seams of every pocket, it isn’t such a leap to agree that anything can be expressed in information.

That’s the idea some theorists are pursuing as they search for ever-more elegant and concise descriptions of the laws that govern our universe. Could our universe, in all its richness and diversity, really be just a bunch of bits?

To some theorists, though, information is more than just a description of our universe and the stuff in it: it is the most basic currency of existence, occupying what theorist Paul Davies terms the “ontological basement” of reality.

The rules of quantum information provide the most “compact” description of physics, says Vlatko Vedral, professor of quantum information theory at the University of Oxford and the National University of Singapore. “Information, it seems to me, requires fewer assumptions about anything else we could postulate. As soon as you talk about matter and energy, you have to write down the laws that govern matter and energy.”

Does this mean that our universe is made of information, as some headlines claim?

“It strikes me as a contentless question,” says Aaronson. “To say that matter and energy are important in physics is to say something with content.” You can imagine a universe barren of matter and energy, after all; specifying that our universe is furnished with both tells you something about it and distinguishes it from other possible universes. “But I don’t know how you could even conceive of a universe” without information, he says.

Yet, as a fresh way of thinking about, well, what the universe is about, information has touched off provocative work in computer science and theoretical astrophysics, apparently disparate fields that may share a deep link manifested by that cosmic Rosetta stone, the black hole. But before we dive into the black hole, let’s step back to take a deeper look at information itself.

All messages contain information, but not all messages are created equal. “Unexpected things have high information content,” says Vedral. Take a sunrise, for example. “If the sun rises tomorrow, you won’t see any newspaper writing about it. But of course if it didn’t rise, it would be a major event.”

We sense intuitively that “surprises,” like a missed sunrise, carry more information than routine events. Claude Shannon, widely considered the father of information theory, formalized this intuition by defining a quantity that’s now known as “Shannon entropy.” The Shannon entropy of a message is related to the sum of the logarithm of the probability of each bit taking on a particular value. That’s a mouthful, but, Vedral explains, it mathematically captures two important features of information: the value of surprises, and the fact that information is “additive”—that is, that the total information contained in two, three, four, or a billion unrelated events is equal to the sum of the information in each one.

Physicists describe entropy a little differently, often speaking in terms of the “disorder” of a system. More precisely, entropy is the number of different ways you can rearrange the littlest parts of a system and still get the same big system. A bucket full of red Legos, for instance, has high entropy. Shake it up, spin it around, and you still have what you began with: a bucket of red Legos. Assemble those same blocks into a Lego castle, though, and you’ve slashed the entropy; moving a single block nets you a different “macroscopic” system.

Whichever perspective you choose, the essential result is the same. Take the paragraph you’re reading right now, for instance, with its many different letters, punctuation marks, and spaces arranged in a very particular order. It contains more information, and thus has higher entropy, than a paragraph like this one—

aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa

aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa

aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa

aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa

aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa

aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa

aaaaa

—even though they contain the same number of characters. Entropy, then, provides a measure of not just the length but the information content of a message.

Now imagine trying to build the ultimate hard drive, one that holds the maximum amount of information allowed by physics. Why should physics place a limit on the information storage capacity of this hypothetical hard drive? Thinking it over from a purely classical perspective, it seems that you could store an infinite amount of information. But when we add quantum mechanics to the mix, we introduce fundamental limits on the accuracy of our measurements. These limits cause entropy to max out at about 10 69 bits per square meter. “If you tried to pack information more densely than that,” says Aaronson, “your hard drive would collapse into a black hole.” That’s not just a whimsical footnote. Black holes, it turns out, are the universe’s very best information repositories. (Of course, Aaronson points out, they don’t make very practical hard drives, unless you’re willing to wait about 10 70 years to retrieve your data.)

But there’s something odd about the way the entropy of a black hole grows. As physicists Stephen Hawking and Jacob Bekenstein discovered in the 1970s, the entropy of a black hole increases with the black hole’s two-dimensional surface area, as defined by an imaginary spherical shell with radius R s . This is bizarre; you would expect the amount of information you can pack into any object, like a book or a hard drive, to grow with the three-dimensional volume of the object, not its surface area.

This discrepancy is more than just theoretical arcana. To physicists, it suggests that the fundamental laws of physics may have a simpler representation in two dimensions rather than the traditional three. In 1997, the Argentinian physicist Juan Maldacena, now at the Institute for Advanced Study, took advantage of this idea to work out a mathematical “duality” between our universe and one with fewer dimensions, one time dimension, and no gravity. This provides a handy mathematical back door; problems that are difficult to solve in one domain may shake out easily in the other.