Big Think / YouTube

Moore’s Law is finally breaking down, according to theoretical physicist Michio Kaku. He’s talking about the so-called law that says the number of transistors that can be fit on a computer chip will double every two years, resulting in periodic increases in computing power.

According to Kaku:

…in about ten years or so, we will see the collapse of Moore’s Law. In fact, already, already we see a slowing down of Moore’s Law. Computer power simply cannot maintain its rapid exponential rise using standard silicon technology. Intel Corporation has admitted this.

It’s true. At the International Supercomputing Conference 2011 last June, Intel architecture group VP Kirk Skaugen said something about Moore’s Law not being sufficient, by itself, for the company to ramp up to exascale performance by 2018. But he went on to tout Intel’s tri-gate technology (the company’s so-called “3D” processors) as the solution, which Skaugen claimed translates to “no more end of life for Moore’s Law.”

(MORE: Meet the Quantum Computer Inside a Diamond — Does It Run ‘Forever’?)

Moore’s Law, introduced by Intel co-founder Gordon Moore in a 1965 paper, was never a law in any scientific sense — it’s always been more a rule of thumb (that, and “Moore’s Rule” sounds so much less authoritative). And as others have pointed out, given Intel’s dominance in the chip industry for much of the period in which Moore’s Law has applied (or appeared to), there’s a self-fulfilling prophecy angle in which the dominant industry player sets the pace for its own benefit.

(VIDEO: The Singularity: How Scared Should We Be?)

Moore also clarified in a 2003 interview that the idea of computer power doubling every 18 months — sometimes mistaken as the basis of Moore’s Law — was advanced by Intel’s David House. While such performance gains could be achieved by Moore’s prediction that transistor counts would double every two years, House apparently calculated that transistors would get faster as well, resulting in computing performance doubling every 18 months (in a later 2005 interview, Moore admitted “we’re doing a little better than [24 months]”).

Despite Intel’s recent advances with tri-gate processors, Kaku argues the company has merely prolonged the inevitable: the law’s collapse due to heat and leakage issues.

“So there is an ultimate limit set by the laws of thermal dynamics and set by the laws of quantum mechanics as to how much computing power you can do with silicon,” says Kaku, noting “That’s the reason why the age of silicon will eventually come to a close,” and arguing that Moore’s Law could “flatten out completely” by 2022.

Where do we go once Gordon Moore’s axiom runs out of steam? Kaku hypothesizes several options: protein computers, DNA computers, optical computers, quantum computers and molecular computers. And then he makes a bet:

If I were to put money on the table I would say that in the next ten years as Moore’s Law slows down, we will tweak it. We will tweak it with three-dimensional chips, maybe optical chips, tweak it with known technology pushing the limits, squeezing what we can.

Kaku then invokes parallelism as another stop-gap measure, a concept that’s been around for decades, but assuming the exponential requirements for processing power hold, “Sooner or later even three-dimensional chips, even parallel processing, will be exhausted and we’ll have to go to the post-silicon era,” says Kaku.

(MORE: Meet Intel’s Crazy 50-Core ‘Knights Corner’, World’s First 1 TFLOPS Processor)

How would a molecular computer work? Imagine molecules in the shape of a valve, says Kaku.

You turn the valve one way and the electricity stops through that molecule. You turn it the other way and electricity flows through that molecule just like a pipe and a valve because that’s what a transistor is, a switch, except this switch is molecular rather than a switch made out of piping.

But molecular computing has mass production issues because — surprise! — molecules are teeny-tiny. Why in the world, then, would Kaku invoke even smaller particle-based computers as a viable alternative?

Because quantum computing could produce the “ultimate computer.” Kaku doesn’t explain why, but I’ll summarize: Digital computers set bits to either “0” or “1,” but in a quantum computer, the bits can be “0” and “1,” at the same time, allowing for incredibly fast calculations according to a principle called “superposition.” The problem, and you knew there’d be one, is something called “decoherence.” Kaku explains:

Let’s say I have two atoms and they vibrate in unison. If I have two atoms and they vibrate in unison I can shine a light wave and flip one over and do a calculation, but they have to first start vibrating in unison. Eventually an airplane goes over. Eventually a child walks in front of your apparatus. Eventually somebody coughs and then all of the sudden they’re no longer in synchronization. It gets contaminated by disturbances from the outside world. Once you lose the coherence, the computer is useless.

Given that, Kaku says that when Moore’s Law finally collapses by the end of the next decade, we’ll “simply tweak [it] a bit with chip-like computers in three dimensions.” Beyond that, he says “we may have to go to molecular computers and perhaps late in the 21st century quantum computers.”

[via Geek.com]

PHOTOS: A Brief History of the Computer