What are you looking for ?
Advertise with us
Advertise with us

SC24: SK hynix Showcasing Advanced Memory Solutions for AI and HPC

Including CMM-DDR5 (CXL Memory Module), AiMX accelerator card, and Niagara 2.0

At SC24 in Atlanta, GA., SK hynix, Inc. is showcasing its advanced memory solutions for AI and HPC.

Sk Hynix Sc24 01

This annual event, organized by the Association for Computing Machinery and the IEEE Computer Society since 1988, features the latest developments in HPC, networking, storage, and data analysis.

Returning for its 2nd year, SK hynix is underlining its AI memory leadership through a display of memory products and insightful presentations on AI and HPC technologies. In line with the conference’s ‘HPC Creates’ theme which underscores the impact of supercomputing across various industries, the company is showing how its memory solutions drive progress in diverse fields.

Showcasing advanced memory solutions for AI and HPC

Sk Hynix Sc24 03b

At its booth, the company is demonstrating and displaying a range of products tailored for AI and HPC. The products being demonstrated include its CMM (CXL (1) Memory Module)-DDR5 (2), AiMX (3) accelerator card, and Niagara 2.0 among others.

Sk Hynix Sc24 04

The live demonstration of CMM-DDR5 with a server platform featuring Xeon 6 processors shows how CXL memory technology accelerates AI workloads under various usage models. Moreover, visitors to the booth can learn about the latest CMM-DDR5 product with EDSFF (4) which offers improvements in TCO (5) and performance.

Sk Hynix Sc24 06

Another live demonstration features AiMX integrated in an ASRock Rack Server to run Meta’s Llama 3 70B, a large language model (LLM) with 70 billion parameters. This demonstration highlights AiMX’s efficiency in processing large datasets while achieving high performance and low power consumption, addressing the computational load challenges posed by attention layers (6) in LLMs.

Sk Hynix Sc24 08

Among the other technologies being demonstrated is Niagara 2.0. The CXL pooled memory solution enables data sharing to minimize GPU memory shortages during AI inference (7), making it for LLM models. The company is also demonstrating an HBM with near-memory processing (NMP) (8) which accelerates indirect memory access (9), a frequent occurrence in HPC. Developed with Los Alamos National Laboratory (LANL), the solution highlights the potential of NMP-enabled HBM to advance next-gen technologies.

 

Sk Hynix Sc24 05Another demonstration is showcasing SK hynix’s updated OCS (10) solution, which offers significant improvements in analytical performance for real-world HPC workloads compared to the iteration displayed at SC23. Co-developed with LANL, OCS addresses performance issues in traditional HPC systems by enabling storage to independently analyze data, reducing unnecessary data movement and improving resource efficiency. Additionally, the company is demonstrating a checkpoint offloading SSD (11) prototype that improves LLM training resource utilization by enhancing performance and scalability.

Sk Hynix Sc24 14In addition to running product demonstrations, SK hynix is also displaying a robust lineup of data center solutions, including its HBM3E (12). The 5thGen HBM provides high-speed data processing, optimal heat dissipation, and high capacity, making it essential for AI applications. Alongside HBM3E are the company’s rapid DDR5 RDIMM and MCR DIMM products, which are tailored for AI computing in high-performance servers. Enterprise SSDs (eSSDs) including the Gen 5 PS1010 and PEB110 are also on display. Offering ultra-fast RW speeds, these SSD solutions are vital for accelerating AI training and inference in large-scale environments.

Highlighting potential of memory through expert presentations

Jongryool Kim presenting on advancements in memory and storage for HPC and AI systems

Sk Hynix Sc24 15During the conference, Jongryool Kim, research director, AI System Infra, presented on Memory & Storage: The Power of HPC/AI, highlighting the memory needs for HPC and AI systems. He focused on 2 key advancements including near-data processing technology using CXL, HBM, and SSDs to improve performance, and CXL pooled memory for better data sharing across systems.

Jeoungahn Park delivering presentation on OCS

Sk Hynix Sc24 16

Jeoungahn Park, technical leader, sustainable computing team also took to the stage for his talk on Leveraging Open Standardized OCS to Boost HPC Data Analytics. Park explained how OCS enables storage to automatically recognize and analyze data, thereby accelerating data analysis in HPC. He added how OCS enhances resource efficiency and integrates with existing analytics systems, as well as how its analysis performance has been verified in real-world HPC applications.

At SC24, SK hynix is solidifying its status as a pioneer in memory solutions which are driving innovations in AI and HPC technologies. Looking ahead, the company will continue to push technological boundaries with support from its partners to shape the future of AI and HPC.

(1) Compute Express Link (CXL): A PCIe-based next-generation interconnect protocol on which high-performance computing systems are based.
(2) CXL Memory Module-DDR5 (CMM-DDR5): A next-gen DDR5 memory module utilizing CXL technology to boost bandwidth and performance for AI, cloud, and high-performance computing.
(3) Accelerator-in-Memory Based Accelerator (AiMX): SK hynix’s specialized accelerator card tailored for large language model processing using GDDR6-AiM chips.
(4) Enterprise and Data Center Standard Form Factor (EDSFF): A collection of SSD form factors specifically used for data center servers.
(5) Total cost of ownership (TCO): The complete cost of acquiring, operating, and maintaining an asset, including purchase, energy, and maintenance expenses.
(6)Attention layer: A mechanism that enables a model to assess the relevance of input data, prioritizing more important information for processing.
(7)AI inference: The process of using a trained AI model to analyze live data for predictions or task completions.
(8) Near-memory processing (NMP): A technique that performs computations near data storage, reducing latency and boosting performance in high-bandwidth tasks like AI and HPC.
(9) Indirect memory access: A computing addressing method in which an instruction providing the address of a memory location that contains the actual address of the desired data or instruction.
(10) Object-based computational storage (OCS): A storage architecture that integrates computation within the storage system, enabling local data processing and minimizing movement to enhance analytical efficiency.
(11) Checkpoint offloading SSD: A storage solution that stores intermediate data during AI training, improving efficiency and reducing training time.
(12) HBM3E: The 5th-gen High Bandwidth Memory (HBM), a high-value, high-performance product that revolutionizes data processing speeds by connecting multiple DRAM chips with through-silicon via (TSV).

Read also :
Articles_bottom
ExaGrid
AIC
ATTOtarget="_blank"
OPEN-E