SK Hynix Launches Mass Production of Advanced 12-Layer HBM3E Memory


September 27, 2024 by our News Team

SK hynix has launched mass production of the first-ever 12-layer HBM3E memory module with 36 GB capacity, setting a new standard in high-performance computing and solidifying their position as a leader in the AI memory market.

1. Massive 36 GB capacity - SK hynix's new HBM3E memory module boasts the largest capacity available today, making it a game-changer for high-performance computing. 2. Advanced MR-MUF process - SK hynix's innovative engineering techniques not only increase capacity, but also improve heat dissipation and stability. 3. Fastest memory speed on the market - With a speed of 9.6 Gbps, this new HBM3E module is the fastest on the market, making it a powerful tool for data-heavy tasks like AI and graphics processing.


In a move that’s set to shake up the world of high-performance computing, SK hynix has just kicked off mass production of what it claims is the first-ever 12-layer HBM3E memory module, boasting a whopping 36 GB of capacity. This is the largest HBM (High Bandwidth Memory) product available today, and it’s a significant leap from the previous 8-layer models that were rolled out just six months ago. It’s a bit like watching a tech race where the finish line keeps moving further away—just when you think you’ve seen the peak of innovation, someone like SK hynix comes along and raises the bar yet again.

For those who might not be deep in the tech weeds, HBM is a type of memory that’s designed to handle the intense demands of data-heavy tasks, like artificial intelligence (AI) and graphics processing. Think of it as the high-octane fuel for your computer’s brain. SK hynix has been in the HBM game since 2013, and they’ve been steadily pushing the envelope ever since. From HBM1 to the newly minted HBM3E, they’ve been the only player to offer a complete lineup. It’s a bit like being the only chef in a culinary competition who can whip up every dish on the menu.

What’s particularly interesting about this new 12-layer product is how SK hynix managed to cram more memory into a space that’s just as thick as its previous offerings. By stacking 12 layers of 3 GB DRAM chips, they’ve increased capacity by 50%. They did this by making each chip 40% thinner, which sounds impressive but also raises a few eyebrows. Thinner chips could lead to structural issues, right? Well, SK hynix tackled that with their Advanced MR-MUF process, which not only improves heat dissipation by 10% but also enhances stability. It’s almost like they’ve built a skyscraper that can withstand strong winds by using cutting-edge engineering techniques.

Now, let’s talk numbers for a second. The memory speed has been cranked up to 9.6 Gbps, which is the fastest on the market. To put that into perspective, if you were to run a Large Language Model (LLM) like Llama 3 70B on a single GPU equipped with four of these HBM3E modules, it could read a staggering 70 billion parameters 35 times in just one second. That’s a data crunching capability that could leave even the most seasoned tech enthusiasts in awe.

Justin Kim, the head of AI Infra at SK hynix, expressed his excitement about this breakthrough, emphasizing the company’s commitment to leading the AI memory market. He mentioned the challenges posed by the rapidly evolving demands of AI technology and how they’re gearing up to meet those needs head-on. It’s a bold claim, but in an industry that’s constantly evolving, staying ahead of the curve is no small feat.

As we continue to witness the rapid advancements in AI and computing technology, it’s clear that companies like SK hynix are not just keeping pace; they’re setting the tempo. With this new 12-layer HBM3E, they’re not only addressing the current demands of AI companies but also paving the way for future innovations. It’ll be fascinating to see how this technology will be utilized in real-world applications, and whether it truly meets the lofty expectations set by its creators. After all, in the fast-paced tech world, the only constant is change. What are your thoughts? Are we ready for this next leap in memory technology?

SK Hynix Launches Mass Production of Advanced 12-Layer HBM3E Memory

About Our Team

Our team comprises industry insiders with extensive experience in computers, semiconductors, games, and consumer electronics. With decades of collective experience, we’re committed to delivering timely, accurate, and engaging news content to our readers.

Background Information


About SK hynix: SK Hynix is a prominent South Korean semiconductor company known for its innovative contributions to the global technology landscape. Specializing in the production of memory solutions, SK Hynix has played a vital role in shaping the semiconductor industry. With a commitment to research and development, they have continuously pushed the boundaries of memory technology, resulting in cutting-edge products that power various devices and applications.

SK hynix website  SK hynix LinkedIn

Technology Explained


GPU: GPU stands for Graphics Processing Unit and is a specialized type of processor designed to handle graphics-intensive tasks. It is used in the computer industry to render images, videos, and 3D graphics. GPUs are used in gaming consoles, PCs, and mobile devices to provide a smooth and immersive gaming experience. They are also used in the medical field to create 3D models of organs and tissues, and in the automotive industry to create virtual prototypes of cars. GPUs are also used in the field of artificial intelligence to process large amounts of data and create complex models. GPUs are becoming increasingly important in the computer industry as they are able to process large amounts of data quickly and efficiently.


HBM3E: HBM3E is the latest generation of high-bandwidth memory (HBM), a type of DRAM that is designed for artificial intelligence (AI) applications. HBM3E offers faster data transfer rates, higher density, and lower power consumption than previous HBM versions. HBM3E is developed by SK Hynix, a South Korean chipmaker, and is expected to enter mass production in 2024. HBM3E can achieve a speed of 1.15 TB/s and a capacity of 64 GB per stack. HBM3E is suitable for AI systems that require large amounts of data processing, such as deep learning, machine learning, and computer vision.


LLM: A Large Language Model (LLM) is a highly advanced artificial intelligence system, often based on complex architectures like GPT-3.5, designed to comprehend and produce human-like text on a massive scale. LLMs possess exceptional capabilities in various natural language understanding and generation tasks, including answering questions, generating creative content, and delivering context-aware responses to textual inputs. These models undergo extensive training on vast datasets to grasp the nuances of language, making them invaluable tools for applications like chatbots, content generation, and language translation.





Leave a Reply