Nvidia bridge 2 shot
The H100 NVL has a full 6144-bit memory interface (1024-bit for each HBM3 stack) and memory speed up to 5.1 Gbps. This means that the maximum throughput is 7.8GB/s, more than twice as much as the H100 SXM. Large Language Models require large buffers and higher bandwidth will certainly have an impact as well. NVIDIA H100 NVL for Large Language Model Deployment is ideal for deploying massive LLMs like ChatGPT at scale. The new H100 NVL with 96GB of memory with Transformer Engine acceleration delivers up to 12x faster inference performance at GPT-3 compared to the prior generation A100 at data center scale.
Graphic Brand: Nvidia |
Graphic Series: A100 |
Graphic GPU Name: GH100 |
Graphic Memory type: HBM2 |
Graphic Bus: PCIe Gen 4.0 x16 |
Graphic Display Connectors: None |
Graphic Power consumption (TDP): 400W |
Graphic Supplementary power connectors: 2 x 8-pin PCIe |