Intel Vision: Liqid, Intel, and IntelliProp Deliver Dynamic Memory With CXL Demo
Using CXL-enabled fabric adapters and fabric-attached memory, companies showcase DRAM pooling, sharing, and composability.
This is a Press Release edited by StorageNewsletter.com on May 16, 2022 at 2:01 pmLiqid, Inc. announced continued success in the area of composable memory with another Matrix software demonstration of DRAM memory attached via the Compute Express Link (CXL) protocol.
The company worked with Intel Corp. and memory and storage solutions provider IntelliProp, Inc., utilizing CXL to compose DRAM to hosts via software in tandem with other data center elements such as GPU, FPGA, and NVMe storage. With the ability to disaggregate memory from the CPU for the first time, CXL promises the kind of performance and flexibility necessary to address the crushing data requirements of AI+ML, and other high-value applications increasingly weaving their way into every element of the connected world.
“Composable memory via CXL allows us to disaggregate the last remaining component from the CPU, enabling new levels of efficiency, flexibility, and agility for memory intensive applications, just like we have been doing for GPU, NVMe storage, and FPGA for years,” said Ben Bolles, executive director, product management, Liqid. “We are pleased to work with industry leaders like Intel and the groundbreaking innovators at IntelliProp to pioneer new levels of performance and flexibility with these early CXL proofs of concept as the need to compose DRAM in tandem with other data center elements becomes critically important to the day-to-day operations of industry, government, academia, and wide range of other data-centric verticals.”
Based on the coming PCIe Gen-5 standard, CXL delivers high-speed CPU-to-memory connections, allowing for the decoupling of DRAM from the CPU, the final hardware element to be disaggregated. CXL is a high-speed interconnect delivering performance to support the pooling and sharing of emerging memory solutions based on the DDR5 memory standard, which increases the density and bandwidth of DRAM while reducing power consumption requirements, thus increasing overall data center efficiency.
As a wave of CXL-supported servers becomes commercially available, composability incorporated into any server refresh will both enable existing resources to continue to be utilized, while also deploying DRAM as a shared, pooled, bare-metal resource that can be deployed in tandem with accelerator technologies already available to the firm’s Composable Disaggregated Infrastructure (CDI) software. With native support for CXL, the Matrix software can pool and compose DDR5 memory in tandem with GPU, NVMe, persistent memory, FPGA, and other accelerator devices.
Intel is a pioneer in DDR5 memory and a founding member of the CXL consortium. Servers based on 4thGen Xeon Scalable processors, formerly codenamed Sapphire Rapids, feature integrated support for CXL, DDR5, and PCIe Gen-5 for performance improvements across a range of workloads and with Intel Advanced Matrix Extensions (AMX) that achieve up to an 8x performance increase for AI inference1[2].
“New industry standards such as CXL enable new memory and server architectures which address the evolving requirements of modern computing,” said Jim Pappas, director, Technology Initiatives, Intel. “We applaud the investment of companies like Liqid and IntelliProp which provide new proofs of concept that give IT professionals an early look at the potential of compute ecosystems based on CXL technology.”
The IntelliProp DL Memory Mesh is created using CXL fabric adapters and fabric attached memory which allows sharing of ML training data across multiple CPU nodes reducing cost and power while improving the scale and overall performance of the Deep Learning applications.
“With the advent of CXL, advancements in memory and server technologies from industry stalwarts like Intel, and composable software from Liqid, we finally have the ability to use fabric-attached memory to provide large and multi-server accessible memory pools when, where, and at the capacity that is required,” said Hiren Patel, CEO, IntelliProp. “We’re excited to offer this joint demonstration of the power of CXL with Intel and Liqid, and to work with leaders in data center innovation as we finally approach the kind of memory-intensive compute that will support and accelerate advancements in higher-order applications that are being driven by AI+ML.”
1[2] Source: Results may vary.