NVIDIA’s AI Enterprise software shown at Supercomputing ‘23 connects accelerated computing to large language model use cases.
At the Supercomputing ‘23 conference in Denver on Nov. 13, NVIDIA announced expanded availability of the NVIDIA GH200 Grace Hopper Superchip for high-performance computing and HGX H200 Systems and Cloud Instances for AI training.
Jump to:
The HGX GH200 supercomputing platform, which is built on the NVIDIA H200 Tensor Core GPU, will be available through server manufacturers and hardware providers that have partnered with NVIDIA. The HGX GH200 is expected to start shipping from cloud providers and manufacturers in Q2 2024.
Amazon Web Services, Google Cloud, Microsoft Azure, CoreWeave, Lambda, Vultr and Oracle Cloud Infrastructure will offer H200-based instances in 2024.
NVIDIA HGX H200 features the following:
“To create intelligence with generative AI and HPC applications, vast amounts of data must be efficiently processed at high speed using large, fast GPU memory,” said Ian Buck, vice president of hyperscale and HPC at NVIDIA, in a press release.
NVIDIA will now offer HPE Cray EX2500 supercomputers with the GH200 chip (Figure A) for enhanced supercomputing and AI training. HPE announced a supercomputing solution for generative AI made up in part of NVIDIA’s HPE Cray EX2500 supercomputer configuration.
Figure A
The GH200 includes Arm-based NVIDIA Grace CPU and Hopper GPU architectures using NVIDIA NVLink-C2C interconnect technology. The GH200 will be packaged inside systems from Dell Technologies, Eviden, Hewlett Packard Enterprise, Lenovo, QCT and Supermicro, NVIDIA announced at Supercomputing ’23.
SEE: NVIDIA announced AI training-as-a-service in July (TechRepublic)
“Organizations are rapidly adopting generative AI to accelerate business transformations and technological breakthroughs,” said Justin Hotard, executive vice president and general manager of HPC, AI and Labs at HPE, in a blog post. “Working with NVIDIA, we’re excited to deliver a full supercomputing solution for generative AI, powered by technologies like Grace Hopper, which will make it easy for customers to accelerate large-scale AI model training and tuning at new levels of efficiency.”
Projects like HPE’s show that supercomputing has applications for generative AI training, which could be used in enterprise computing. The GH200 interoperates with the NVIDIA AI Enterprise suite of software for workloads such as speech, recommender systems and hyperscale inference. It could be used in conjunction with an enterprise’s data to run large language models trained on the enterprise’s data.
NVIDIA announced partnerships with supercomputing centers around the world. Germany’s Jülich Supercomputing Centre’s scientific supercomputer, JUPITER, will use GH200 superchips. JUPITER will be used to create AI foundation models for climate and weather research, material science, drug discovery, industrial engineering and quantum computing for the scientific community. The Texas Advanced Computing Center’s Vista supercomputer and the University of Bristol’s upcoming Isambard-AI supercomputer will also use GH200 superchips.
Cloud providers Lambda and Vultr offer NVIDIA GH200 in early access now. Oracle Cloud Infrastructure and CoreWeave plan to offer NVIDIA GH200 instances in the future, starting in Q1 2024 for CoreWeave; Oracle did not specify a date.