ADVERTISEMENT

The Next 20 Years of Microchips: Pushing Performance Boundaries

Designers are pushing all the boundaries to make integrated circuits smaller, faster and cheaper
THIS IS A PREVIEW.
or subscribe to access the full article.



Courtesy of AMD Global Communications

 In 1975 electronics pioneer Gordon Moore famously predicted that the complexity of integrated-circuit chips would double every two years. Manufacturing advances would allow the chip’s transistors to shrink and shrink, so electrical signals would have to travel less distance to process information. To the electronics industry and to consumers, Moore’s Law, as it became known, meant computerized devices would relentlessly become smaller, faster and cheaper. Thanks to ceaseless innovation in semiconductor design and fabrication, chips have followed remarkably close to that trajectory for 35 years.

Engineers knew, however, they would hit a wall at some point. Transistors would become only tens of atoms thick. At that scale, basic laws of physics would impose limits. Even before the wall was hit, two practical problems were likely to arise. Placing transistors so small and close together while still getting a high yield—usable chips versus defective ones—could become overly expensive. And the heat generated by the thicket of transistors switching on and off could climb enough to start cooking the elements themselves.

THIS IS A PREVIEW.
or subscribe to access the full article.
Digital Issue $7.99
Digital Issue + All Access Subscription $99.99 Subscribe
Rights & Permissions
Share this Article:

Comments

You must sign in or register as a ScientificAmerican.com member to submit a comment.
Scientific American Holiday Sale

Scientific American Mind Digital

Get 6 bi-monthly digital issues
+ 1yr of archive access for just $9.99

Hurry this offer ends soon! >

X

Email this Article

X