“Those who do not know their history are doomed to repeat it.” With the phenomenal success of the transistor’s first 75 years, it feels like maybe we should both know and repeat. That seems to have been the reaction you (dear devoted IEEE Spectrum reader) took with your semiconductor article consumption. You dove deep with us into the inner workings of the very first transistor, but you also wanted to know what comes next. In between, taking some time to appreciate how fundamental forces are smashing up big system-on-chips, what an injection of government money could do to chip manufacturing in the United States, and some stranger stuff.
So to recap the semiconductor stories with the highest traffic numbers from this year, we’ve put together this set of highlights:
It’s hard to overstate—considering the utter ubiquity and profound significance of electronic technology in people’s lives today—the importance of the transistor, which was invented in December 1947. After all, we likely produced 2 billion trillion of them this year alone. Yet there was only one in 1947, and it worked pretty weirdly. How we got from 1 to 2 billion trillion per year is a separate story. But what Spectrum readers were most interested in was how we’re going to make future billion trillions. The next step (well, the next, next step for some companies) seems pretty clear; it’s the complementary FET. That’s a PMOS and NMOS transistor built atop each other, likely all at once, potentially cutting the size of logic cells in half. What comes after that? At the 100th anniversary of the transistor there will likely be variations on that theme, our experts say.
I’ve been listening to promises of cheap, flexible electronics my whole career. These could add intelligence to every currently-dumb object on store shelves today. So where’s my smart-banana? This year, we finally got an answer: Until now, nobody has built working plastic processors that can be made in the billions for less than a penny each. Even the simplest industry-standard microcontrollers are too complex to make on plastic in bulk, because the yield of good “chips” (for want of a better word) is super low. Engineers in Illinois and England came to the rescue with an architecture simple enough to breach the one-penny barrier.
In an interview in early December, Ann B. Kelleher, general manager of technology development at Intel really tried to make me understand how a concept called system technology co-optimization (STCO) was going save Moore’s Law. It’s not that I wasn’t buying it; I was. (And Spectrum has been writing about STCO in one way or another for nearlytwo decades.) It’s just that it’s an idea that really needs an example to go with it. In STCO, you start with the software workload, figure out the functions needed for that, and then determine which functions should be implemented using which semiconductor manufacturing technology. All of those things feed back on each other, and so must be “co-optimized.” It’s possible to do that, because you can now break a system-on-chip into functional chiplets and stitch them together using 3D and other advanced packaging, so they act like a single large chip. As for the example, Kelleher recommends looking at the Ponte Vecchio accelerator behind the Aurora supercomputer. It’s made up of 47 pieces of silicon made using four different processes, three of which weren’t even Intel’s.
This was the year 3D chip packaging showed up in force, at least in the high-end logic arena. AI supercomputer startup Graphcore used some TSMC 3D chip-stacking tech to improve the flow of power to its chip in a way that could lead to a 40 percent speed up in training neural networks. AMD used a different TSMC tech to give it’s compute chiplets a major memory boost. And Intel through every chip packaging tech they had to Ponte Vecchio to build a 47-chiplet monster. As you’ve noticed if you’ve read any of the other items in this post, this is the way of the future.
Legislation aimed at increasing semiconductor manufacturing in the United States finally became law after a multiyear journey that saw many mutations and delays. The CHIPS and Science Act, part of a larger $280 billion package, provides about US $52 billion for new or expanded facilities that make semiconductors or chipmaking equipment. And it arrived amidst efforts by other nations and regions to boost chip manufacturing, an industry increasingly seen as a key to economic and military security. Now, the hard part—figuring out who gets how much money.
Scientists have been chasing electronic devices based on topological materials, which basically protect the flow of current from disturbance because of the same math that tells you a donut and a coffee mug are fundamentally the same shape but cereal bowl and coffee mug are not. Harvard scientists took a detour they hope will lead to some breakthroughs, by building topological transistors that modulate sound instead of electrons. The result is a weird looking honeycomb system, that looks like it belongs in the inner workings of Iron Man’s suit.
In 2022, the open source instruction set architecture RISC-V began what analysts see as a rapid penetration of the machine learning market. Early in the year, there was already silicon from startups aimed toward servers, like Esperanto AI, and those targeting “edge” computing like Kneron. Growth should continue to accelerate now that there’s a standard set of vector instructions adopted by the organization governing RISC-V.
Original Source: https://spectrum.ieee.org/3d-chips-plastic-processors