Nvidia H100 NVL 94G (Bulk)
The H100 NVL has a full 6144-bit memory interface (1024-bit for each HBM3 stack) and memory speed up to 5.1 Gbps. This means that the maximum throughput is 7.8GB/s, more than twice as much as the H100 SXM. Large Language Models require large buffers and higher bandwidth will certainly have an impact as well. NVIDIA H100 NVL for Large Language Model Deployment is ideal for deploying massive LLMs like ChatGPT at scale. The new H100 NVL with 96GB of memory with Transformer Engine acceleration delivers up to 12x faster inference performance at GPT-3 compared to the prior generation A100 at data center scale.
Graphic Brand: Nvidia |
Graphic Series: H100 NVL |
Graphic Model: H100 NVL 94GB |
Graphic GPU Name: GH100 |
Graphic CUDA cores: 14,592 |
Graphic Core clock speed GPU: 1785 MHz |
Graphic Core clock speed Memory: 2619 MHz |
Graphic Memory RAM amount: 94 GB |
Graphic Memory type: HBM3 |
Graphic Bus: PCI Express 5.0 x16 (PCIe) / NVLink (SXM) |
Graphic Display Connectors: None |
Graphic Power consumption (TDP): 400W |
Graphic Supplementary power connectors: 1x 16-pin |