AWS and NVIDIA Announce Strategic Collaboration to Offer New Supercomputing Infrastructure, Software, and Services for Generative AI
Retrieved on:
Tuesday, November 28, 2023
Networks, Internet, Hardware, Artificial Intelligence, Data Management, Technology, Semiconductor, Other Technology, Cloud, Titan, Amazon Elastic Compute Cloud, L4, NVLink, Drug discovery, Memory, Robotics, Amazon, Multimedia, Graphics processing unit, Amazon Robotics, Digital, AMZN, AI, Amazon Virtual Private Cloud, Amazon (company), Amazon SageMaker, NVDA, Machine learning, Ford Barra engine, HPC, Simulation, GPU, LLM, Research, Natural language processing, Nvidia Omniverse, FMS, PCI Express, FP8, EFA, Deep learning super sampling, Central processing unit, VPC, DGX, EC2, Software, 3D, AWS, TB, Amazon Elastic Block Store, Cryptocurrency, Video game, Medical device, Hopper, Design, NVIDIA, G6
The NVIDIA GH200 NVL32 multi-node platform connects 32 Grace Hopper Superchips with NVIDIA NVLink and NVSwitch technologies into one instance.
Key Points:
- The NVIDIA GH200 NVL32 multi-node platform connects 32 Grace Hopper Superchips with NVIDIA NVLink and NVSwitch technologies into one instance.
- • NVIDIA and AWS will collaborate to host NVIDIA DGX Cloud , NVIDIA’s AI-training-as-a-service, on AWS.
- DGX Cloud on AWS will accelerate training of cutting-edge generative AI and large language models that can reach beyond 1 trillion parameters.
- NVIDIA DGX Cloud is an AI supercomputing service that gives enterprises fast access to multi-node supercomputing for training the most complex LLM and generative AI models, with integrated NVIDIA AI Enterprise software and direct access to NVIDIA AI experts.