The AI chip industry has a heat problem. Not in the abstract — in the literal, physical sense that is forcing engineers to rethink the most fundamental assumption in computing: that the bottleneck is how fast a chip can calculate. For leading-edge artificial intelligence accelerators, the bottleneck has moved. The limiting factor now is how fast you can move the data between chips — and the industry has found its answer in light.
That answer has a name at the manufacturing level: TSMC COUPE, a silicon photonics platform that combines a 6-nanometer control chip with a 65-nanometer photonic chip in a single bonded package, according to Semiconductor Engineering. It also has two names at the customer level: Nvidia and Broadcom. Both companies have independently selected COUPE — Compact Universal Photonics Engine — as the basis for their next-generation co-packaged optics, or CPO. When two fierce competitors make the same proprietary bet at the same time, that is not a partnership. That is a standard being set in real time.
The clearest proof that the market understands this came last May, when AMD acquired a startup called Enosemi. AMD announced the acquisition on May 28, 2025. The company raised $150,000 in venture capital. It had sixteen employees. It was founded in 2023. AMD paid enough to make it an acqui-hire with photonic IP — the kind of deal where the acquirer is buying something that does not yet have a market price because nobody else saw it coming. The VCs slept through what AMD wanted badly enough to acquire outright.
The physics is not subtle. Enosemi estimated in a white paper that leading-edge high-performance chips now spend as much as half of their total power just moving data between components. Mike Hogan, GlobalFoundries' chief business officer, frames it around four metrics that determine interconnect efficiency: reach, bandwidth density, energy efficiency, and compute utilization. The first three all improve with optical interconnects. The fourth suffers without them. "Scaling east-west traffic requires optical interconnect," Hogan said. He was not wrong.
Light moves data faster and more efficiently than electrons through copper wire. The challenge has been miniaturizing optical components — lasers, modulators, waveguides, photodetectors — to the point where they can live inside a data center rack instead of just between continents. Three integration approaches are competing for that rack space. The first is pluggable modules: relatively large, modular, and easy to replace when they fail. They work well for connections between racks or across a data center floor. The second is co-packaged optics, which integrates optical elements directly with the electronic control circuits they serve, reducing the distance signals have to travel. The third is optical I/O, which combines optical and electronic chips into a single unit that behaves as one device — the most integrated approach, and the most technically demanding.
The efficiency difference is measured in picojoules per bit. Conventional copper interconnects in AI rack-scale systems consume more than 30 picojoules per bit. Co-packaged optics on a substrate brings that down to under 5 picojoules per bit. Optical I/O integrated onto the processor interposer gets you below 2. That is not incremental improvement. For a rack pulling 100 kilowatts — Nvidia's Blackwell architecture in a fully loaded configuration — halving the energy cost of data movement is the difference between a workable power budget and an engineering crisis.
TSMC's COUPE is the manufacturing platform that makes CPO viable at scale. It bonds a 6-nanometer electronic integrated circuit to a 65-nanometer silicon photonic integrated circuit using its SoIC-X hybrid bonding technology, then offers a roadmap toward integration onto the interposer itself through CoWoS — the same advanced packaging platform TSMC used to become indispensable to the AI chip design ecosystem. The numbers from TSMC: 40 percent power reduction at the same speed versus conventional micro-bump approaches, or a 170 percent speed gain at the same power envelope. Neither Nvidia nor Broadcom has found a better answer, which is why both are using it.
The cost structure is shifting accordingly. In Nvidia's own Quantum-X800 switch — the current generation InfiniBand solution — 72 optical engines and 18 external light sources are packaged together. The optical engine alone represents 44 to 45 percent of the total CPO bill of materials. The photonic component is no longer the add-on. It is the expensive part.
Meta provided the most concrete evidence that the technology is ready for production deployment. At the Optical Fiber Communications Conference in March, Meta shared reliability data comparing co-packaged optics to pluggable optical transceivers in its own scale-out network infrastructure. The result: CPO outperformed pluggables. The conventional wisdom that co-packaged optics is too immature for prime time is being answered in live data centers, not in conference keynote slides.
This is where the other fabs face an uncomfortable question. GlobalFoundries has an earlier start, a real installed base, and a monolithic integration story through its Fotonix platform. The company expects silicon photonics revenue to exceed $1 billion by 2030, and its acquisition of AMF is expected to contribute more than $75 million in 2026. Tower Semiconductor reported $228 million in silicon photonics revenue in 2025 — more than double the prior year — with more than 50 active customers. Both are real companies with real traction.
But neither GlobalFoundries nor Tower has TSMC's combination of fine-node electronic integrated circuit capability and CoWoS-class packaging integration. The comparison holds as long as the market remains at the level of standalone optical engines and pluggable modules. The moment the conversation moves to fully integrated CPO — optical engines co-packaged directly with switch ASICs or accelerators at scale — the structural advantage shifts to TSMC. From a supply chain standpoint, whether a company's chips are fabbed at GlobalFoundries or Tower, all roads in photonics integration still lead to TSMC for the packaging step that matters most.
The timing is not accidental. Nvidia's backlog for AI compute reached $1 trillion earlier this year — a figure that reflects inference demand doubling roughly every twelve months as AI evolves from simple query-response to deep reasoning to autonomous agents. Each step up the capability ladder processes more tokens per query and demands faster token delivery to the end user. The compute has to be fed, and the wiring between chips is where the bottleneck landed.
What Nvidia and Broadcom have done, independently and without announcement, is answer the question of who makes the platform. The rest of the industry is now figuring out what that means for them.