SK Hynix ships world’s first 12-layer HBM4 samples early
The S.Korean high-bandwidth memory chip giant originally planned to ship the samples in early 2026
By Mar 19, 2025 (Gmt+09:00)
LG Chem to sell water filter business to Glenwood PE for $692 million


KT&G eyes overseas M&A after rejecting activist fund's offer


Kyobo Life poised to buy Japan’s SBI Group-owned savings bank


StockX in merger talks with Naver’s online reseller Kream


Meritz backs half of ex-manager’s $210 mn hedge fund



South Korean memory chip giant SK Hynix Inc. has begun delivering samples of the world’s most advanced, next-generation high bandwidth memory chip, the 12-layer HBM4, to its key customers ahead of its original schedule, further widening its lead over rivals.
The HBM pioneer, a core supplier to artificial intelligence chip leader Nvidia Corp. and other AI chipmakers, announced on Wednesday that it had delivered the first HBM4 samples to its customers for quality evaluation six months earlier than its original timeline, which was early 2026.
SK Hynix said it plans to get ready for mass production of the 12-layer HBM4 within the second half of this year for the immediate start of its supply.
The faster-than-planned sample supply comes in response to mounting pressure from its customers asking for faster, more efficient memory for next-generation AI applications, especially Nvidia, amid an intensifying AI race.
Last November, Nvidia CEO Jensen Huang asked SK Hynix to supply 12-layer HBM4 chips about six months before the Korean chipmaker’s original schedule.
The sixth-generation high-performance memory chip is expected to be fitted in Nvidia’s next-generation graphics architecture Vera Rubin, the successor of the Blackwell AI chip.
On Tuesday, Huang said at the Nvidia GPU Technology Conference (GTC) 2025 – dubbed the “Super Bowl of AI” – that Blackwell Ultra is slated for the second half of 2025, while the Rubin AI chip is expected to launch in late 2026.

This is also the first HBM chip developed in collaboration between SK Hynix and global top foundry player Taiwan Semiconductor Manufacturing Co. (TSMC).
Last December, the Korean chipmaker announced that it would produce the HBM4 chip on a 3-nanometer node.
It said it would adopt TSMC’s advanced logic process for its HBM4’s base die, which is placed at the bottom of an HBM connected to a graphics processing unit and acts as its brain.
MOST ADVANCED HBM
SK Hynix’s latest innovation boasts the industry’s best capacity and speed, which are essential for AI processors that handle vast amounts of data, said the company.
Its 12-layer HBM4 is the industry’s first HBM capable of processing more than 2 terabytes (TB) of data per second. This means data equivalent to more than 400 full-HD movies (5GB each) can be processed in a second, more than 60% faster than the previous generation, the HBM3E, explained the company.
The Korean chipmaker also adopted the advanced MR-MUF process, which helps prevent chip warpage and maximizes product stability by improving heat dissipation, to achieve a capacity of 36GB, the highest among 12-layer HBM products.

SK Hynix has long been a leader in HBM development, holding an estimated 50% market share in the global HBM chip market.
It succeeded in mass-producing the 8- and 12-layer HBM3E in 2024, becoming the world’s first to supply the 12-high HBM3E to Nvidia.
AT GTC 2025
The Korean chipmaker will showcase its HBM and other memory products for AI data centers, as well as on-device and memory solutions for automotive applications at GTC 2025, a global AI conference taking place March 17-21 in San Jose, CA.
It will present the 12-high HBM3E and SOCAMM2, or small outline compression attached memory module, for AI servers, as well as a model of the 12-high HBM4.

HBM technology, which stacks multiple dynamic random access memory (DRAM) chips vertically to accelerate data transfer speeds, has become a cornerstone of the AI revolution.
Nvidia, the dominant force in AI processing, has leaned heavily on HBM technology to enhance its high-end GPUs, which power large-scale AI models such as OpenAI’s ChatGPT.
Rivals Samsung Electronics Co. and Micron Technology Inc. are also working on next-generation HBM solutions, aiming to capture a share of the growing AI infrastructure market.
Especially, SK Hynix’s cross-town rival Samsung recently outlined aggressive expansion plans for its HBM capacity but it is behind SK Hynix in both technology prowess and schedule.
Write to Chae-Yeon Kim at why29@hankyung.com
Sookyung Seo edited this article.
-
Korean chipmakersSK Hynix to produce HBM4 on 3 nm foundry process in 2025
Dec 03, 2024 (Gmt+09:00)
2 Min read -
Korean chipmakersTesla asks Samsung, SK Hynix to supply HBM4 chip samples
Nov 19, 2024 (Gmt+09:00)
3 Min read -
Korean chipmakersSK Hynix chief says no delay in 12-layer HBM3E supply as demand soars
Oct 23, 2024 (Gmt+09:00)
3 Min read -
Korean chipmakersSK Hynix mass-produces 12-layer HBM3E for Q4 shipment
Sep 26, 2024 (Gmt+09:00)
2 Min read -
Korean chipmakersSamsung Electronics, TSMC tie up for HBM4 AI chip development
Sep 05, 2024 (Gmt+09:00)
3 Min read -
EarningsSK Hynix to supply 12-layer HBM3E to Nvidia in Q4; profit soars in Q2
Jul 25, 2024 (Gmt+09:00)
3 Min read -
Korean chipmakersSK Hynix’s HBM chip orders fully booked; 12-layer HBM3E in Q3: CEO
May 02, 2024 (Gmt+09:00)
5 Min read -
Korean chipmakersSK Hynix, TSMC tie up to stay ahead of Samsung for HBM supremacy
Apr 19, 2024 (Gmt+09:00)
4 Min read -
Korean chipmakersSK Hynix mass-produces HBM3E chip to supply Nvidia
Mar 19, 2024 (Gmt+09:00)
3 Min read -
Korean chipmakersSamsung Elec to launch HBM4 in 2025 to win war in AI sector
Oct 10, 2023 (Gmt+09:00)
2 Min read