imec Advances 3D HBM-on-GPU Thermals Through System-Tech Co-Optimization
December 10, 2025 | ImecEstimated reading time: 3 minutes
At the 2025 IEEE International Electron Devices Meeting (IEDM), imec, world-leading research center in advanced semiconductor technologies, presents the first thermal system-technology co-optimization (STCO) study of 3D HBM-on-GPU (high-bandwidth memory on graphical processing unit), a promising compute architecture for next-gen AI applications. By combining technology and system-level mitigation strategies, peak GPU temperatures could be reduced from 140.7°C to 70.8°C under realistic AI training workloads – on par with current 2.5D integration options. The result demonstrates the strength of combining cross-layer optimization (i.e., co-optimizing the knobs at all the different abstraction layers) with broad technological expertise, a combination that is unique to imec.
Integrating high bandwidth memory (HBM) stacks directly on top of graphical processing units (GPUs) offers an attractive approach for building next-gen compute architectures for data-intensive AI workloads. This 3D HBM-on-GPU promises a huge leap forward in compute density (with four GPUs per package), memory per GPU, and GPU-memory bandwidth compared to current 2.5D integration options where HBM stacks are placed around (one or two) GPUs on a silicon interposer. However, the aggressive 3D integration approach is prone to thermal issues because of higher local power density and vertical thermal resistance.
At 2025 IEDM, imec presents the first comprehensive thermal simulation study of 3D HBM-on-GPU integration that not only identifies thermal bottlenecks but also proposes strategies to increase the architecture’s thermal feasibility. Imec researchers show how co-optimizing technology and system-level thermal mitigation approaches can reduce peak GPU temperatures from 141.7°C to 70.8°C under realistic AI training workloads.
The model assumes four HBM stacks – each consisting of twelve hybrid-bonded DRAM dies – placed directly on top of a GPU using microbumps. Cooling is provided on top of the HBMs. Power maps derived from industry-relevant power profiles are applied to identify local hotspots and compare them to a 2.5D baseline. Without thermal mitigation strategies, the 3D model yields a peak GPU temperature of 141.7°C – far too high for GPU and HBM operation – while the 2.5D integration benchmark peaks at a workable 69.1°C under the same cooling conditions. We used these data as a starting point to evaluate the joint impact of technology and system-level thermal mitigation strategies. Technology-level strategies include, among others, HBM stack merging and thermal silicon optimization. On the system-level side, we assessed the impact of double-sided cooling as well as GPU frequency scaling.
James Myers, System Technology Program Director at imec: “Halving the GPU core frequency brought the peak temperature from 120°C to below 100°C, achieving a key target for the memory operation. Although this step comes with a 28% workload penalty (i.e., a slowdown of AI training steps), the overall package outperforms the 2.5D baseline thanks to a higher throughput density offered by the 3D configuration. We are currently using this approach to study other GPU/HBM configurations (e.g., placing GPUs on top of HBMs), anticipating future thermal constraints.”
Julien Ryckaert, Vice President Logic Technologies at imec: “This is also the first time that we demonstrate the capabilities of imec’s cross-technology co-optimization (XTCO) program in building more thermally robust compute systems. XTCO was launched in 2025 to efficiently align imec’s technology roadmaps with key industry system scaling challenges and is built on four critical system level pillars: compute density, power delivery, thermal, and memory density and bandwidth. It combines our STCO/DTCO mindsets with imec’s broad technology expertise – a unique combination that is of great value in addressing the growth and diversification of compute system demands. We invite companies from within the entire semiconductor ecosystem, including fabless and system companies, to join our XTCO program and collaboratively resolve critical system scaling bottlenecks.”
Testimonial
"The I-Connect007 team is outstanding—kind, responsive, and a true marketing partner. Their design team created fresh, eye-catching ads, and their editorial support polished our content to let our brand shine. Thank you all! "
Sweeney Ng - CEE PCBSuggested Items
Rubin Faces Delays; Blackwell to Drive 70%+ of NVIDIA High-End GPU Shipments in 2026
04/08/2026 | TrendForceAccording to TrendForce’s latest findings on AI servers, NVIDIA’s high-end AI chip shipment mix is expected to change in 2026.
Imec Names NVIDIA’s Jensen Huang as Recipient of the 2026 imec Lifetime of Innovation Award
03/24/2026 | ImecImec, a world-leading research and innovation hub in advanced semiconductor technologies, proudly announces Jensen Huang, founder and CEO of NVIDIA, as the recipient of the 2026 imec Lifetime of Innovation Award.
Top Eight CSPs’ CapEx to Surpass $710B in 2026; Google Leads TPU Deployment
02/25/2026 | TrendForceGlobal CSPs are accelerating investment in AI servers and infrastructure to support expanding AI deployment and upgrades, according to TrendForce’s latest findings on the AI server market.
3Q25 GPU Shipments Up 2.5% as Data Center GPUs Surge 145%
11/25/2025 | JPRJon Peddie Research reports the growth of the global PC-based graphics processor unit (GPU) market reached 76.6 million units in Q3'25 and PC CPU shipments increased to 65.0 million units.
Beyond Design: Micro-ohm Power Delivery Network for AI-driven GPUs
11/18/2025 | Barry Olney -- Column: Beyond DesignThe evolution of modern processors, marked by faster edge transitions, reduced output impedance, and increasingly complex bus architectures, has significantly augmented the demands on PCB infrastructure. These challenges are compounded by AI-driven graphics processing units (GPUs), which require exceptionally high-power delivery at ultra-low operating voltages, placing greater stress on power integrity and layout design.