MWC Las Vegas: Supermicro Introduced SYS-322GB-NR 3U Server Versatile System Design for AI
Features up to 6TB of RDIMM memory and up to 14 E1.S or 6 U.2 NVMe drives, and supports up to 18 GPUs and features dual Xeon 6900 series processors with P-cores.
This is a Press Release edited by StorageNewsletter.com on October 17, 2024 at 2:01 pmSuper Micro Computer, Inc. launches a versatile, high-density infrastructure platform optimized for AI inferencing at the network edge.
As companies seek to embrace complex LLM in their daily operations, there is a need for new hardware capable of inferencing high volumes of data in edge locations with minimal latency. The company’s system combines versatility, performance, and thermal efficiency to deliver up to 10 double-width GPUs in a single system capable of running in traditional air-cooled environments.
“Owing to the system’s optimized thermal design, Supermicro can deliver all this performance in a high-density 3U 20 PCIe system with 256 cores that can be deployed in edge data centers,” said Charles Liang, president and CEO. “As the AI market is growing exponentially, customers need a powerful, versatile solution to inference data to run LLM-based applications on-premises, close to where the data is generated. Our new 3U Edge AI system enables them to run innovative solutions with minimal latency.”
SYS-322GB-NR front
The SYS-322GB-NR includes 2 Xeon 6900 processors with P-cores, 8,800MT/s MRDIMM and up to 20 PCIe 5.0 expansion slots. This system supports a variety of single or double-width GPUs, or to use some of the expansion slots for high-performance I/O or other add-on cards. Additionally, the server features up to 6TB of RDIMM memory and up to 14xE1.S or 6xU.2 NVMe drives.
SYS-322GB-NR rear
One example use case that this system delivers is in the manufacturing industry, where Supermicro’s new system can be deployed on-site at an automated production environment to process data feeds from cameras and sensors without having to transfer the data to a remote location. This capability reduces networking requirements and improves response times. Another environment where the SYS-322GB-NR will excel is large-scale control rooms, where the AI accelerator cards can be partially replaced by multi-display cards to support up to 64 independent displays.
Supermicro at Mobile World Congress (MWC) Las Vegas, NV
The SYS-322GB-NR was on display at the company‘s booth during MWC Las Vegas, NV, October 8-10.
Additionally, the firm will display systems that incorporate NVIDIA Corp., AMD (Advanced Micro Devices Inc.) and Intel Xeon 6 processors, including X14 family edge and telco systems such as:
- SYS-222HE-FTN – the Hyper-E brings data center performance to the telco edge with a dual Xeon 6 processor in a 2U, short-depth form factor with front I/O access
- SYS-212B-FN2T – a 2U short-depth system for AI in telco and edge deployments, featuring a single Xeon 6700 series processor with E-cores and GPU support
- SYS-E403-14B-FRN2T– a box PC sized, wall-mountable edge device capable of bringing the Xeon 6700 series processor with E-cores and GPU support to remote environments
- AS -1115S-FDWTRT – a 1U NEBS compliant system providing Telco performance for ORAN, Core and managed services. This system utilizes the AMD EPYC 8004 Series processor and support of up to 1 single-width GPU accelerator for heavy workloads.
In addition to showcasing the company’s hardware systems, in partnership with NVIDIA, the firm was jointly demonstrate inferencing and AI solutions for on-premise and Edge applications including enterprise AI, retail, telco edge and financial services. The company demonstrated key GenAI solutions that include NVIDIA NIM, NeMo, Metropolis, remote management, security, and networking. For telecom, Supermicro and NVIDIA were demonstrating a live AI RAN solution using the 2 companies solutions showcasing performance, management and AI use cases.
Also, on display at MWC Las Vegas was the new a joint solution from Supermicro and Intel Corp. which combines the ruggedized IP65 Outdoor Edge system with a built-in AI Network Accelerator and Intel Data Center GPU Flex 170. This solution enables fast and cost-effective deployment of multiple private 5G networks as well as Edge AI applications in a single device. The networks can be used and exploited by different users, offering a scalable solution for dense environments such as industrial and campus sites, venues and smart cities.