没有合适的资源?快使用搜索试试~ 我知道了~
联想H100 GPU服务器
资源推荐
资源详情
资源评论
ThinkSystem NVIDIA H100 PCIe Gen5 GPUs
Product Guide
The ThinkSystem NVIDIA H100 GPU delivers unprecedented performance, scalability, and security for every
workload. The GPUs use breakthrough innovations in the NVIDIA Hopper™ architecture to deliver industry-
leading conversational AI, speeding up large language models by 30X over the previous generation.
The NVIDIA H100 GPU features fourth-generation Tensor Cores and the Transformer Engine with FP8
precision, further extending NVIDIA’s market-leading AI leadership with up to 9X faster training and an
incredible 30X inference speedup on large language models. For high-performance computing (HPC)
applications, The GPUs triple the floating-point operations per second (FLOPS) of FP64 and add dynamic
programming (DPX) instructions to deliver up to 7X higher performance.
The following figure shows the ThinkSystem NVIDIA H100 GPU in the double-width PCIe adapter form
factor.
Figure 1. ThinkSystem NVIDIA H100 NVL 94GB PCIe Gen5 Passive GPU
Did you know?
The NVIDIA H100 family is available in both double-wide PCIe adapter form factor and in SXM form factor.
The latter is used in Lenovo's Neptune direct-water-cooled ThinkSystem SD665-N V3 server for the ultimate
in GPU performance and heat management.
The NVIDIA H100 NVL Tensor Core GPU is optimized for Large Language Model (LLM) Inferences, with its
high compute density, high memory bandwidth, high energy efficiency, and unique NVLink architecture.
ThinkSystem NVIDIA H100 PCIe Gen5 GPUs 1
Click here to check for updates
Part number information
The following table shows the part numbers for the ThinkSystem NVIDIA H100 GPU.
Not available in China, Hong Kong and Macau : The H100 GPUs are not available in China, Hong
Kong and Macau.
Table 1. Ordering information
Part number Feature code Description
Double-wide PCIe adapter form factor
4X67A89325 BXAK ThinkSystem NVIDIA H100 NVL 94GB PCIe Gen5 Passive GPU
SXM form factor
CTO only C1HL ThinkSystem NVIDIA HGX H100 80GB 700W 8-GPU Board
CTO only BQQV ThinkSystem NVIDIA H100 SXM5 700W 80G GPU Board
CTO only BUBB ThinkSystem NVIDIA H100 SXM5 700W 94G HBM2e GPU Board
NVLink bridge (for PCIe adapters only, not SXM)
4X67A71309 BG3F ThinkSystem NVIDIA Ampere NVLink 2-Slot Bridge (3 required per pair of
GPUs)
The PCIe option part numbers includes the following:
One GPU with full-height (3U) adapter bracket attached
Documentation
The following figure shows the NVIDIA H100 SXM5 8-GPU Board with heatsinks installed in the
ThinkSystem SR680a V3 and ThinkSystem SR685a V3 servers.
Figure 2. NVIDIA H100 SXM5 8-GPU Board in the ThinkSystem SR680a V3 and SR685a V3 servers
Features
ThinkSystem NVIDIA H100 PCIe Gen5 GPUs 2
Features
The ThinkSystem NVIDIA H100 GPU delivers high performance, scalability, and security for every workload.
The GPU uses breakthrough innovations in the NVIDIA Hopper™ architecture to deliver industry-leading
conversational AI, speeding up large language models (LLMs) by 30X over the previous generation.
The PCIe versions of the NVIDIA H100 GPUs include a five-year software subscription, with enterprise
support, to the NVIDIA AI Enterprise software suite, simplifying AI adoption with the highest performance.
This ensures organizations have access to the AI frameworks and tools they need to build accelerated AI
workflows such as AI chatbots, recommendation engines, vision AI, and more.
The NVIDIA H100 GPU features fourth-generation Tensor Cores and the Transformer Engine with FP8
precision, further extending NVIDIA’s market-leading AI leadership with up to 9X faster training and an
incredible 30X inference speedup on large language models. For high-performance computing (HPC)
applications, the GPU triples the floating-point operations per second (FLOPS) of FP64 and adds dynamic
programming (DPX) instructions to deliver up to 7X higher performance. With second-generation Multi-
Instance GPU (MIG), built-in NVIDIA confidential computing, and NVIDIA NVLink Switch System, the
NVIDIA H100 GPU securely accelerates all workloads for every data center from enterprise to exascale.
Key features of the NVIDIA H100 GPU:
NVIDIA H100 Tensor Core GPU
Built with 80 billion transistors using a cutting-edge TSMC 4N process custom tailored for NVIDIA’s
accelerated compute needs, H100 is the world’s most advanced chip ever built. It features major
advances to accelerate AI, HPC, memory bandwidth, interconnect, and communication at data center
scale.
Transformer Engine
The Transformer Engine uses software and Hopper Tensor Core technology designed to accelerate
training for models built from the world’s most important AI model building block, the transformer.
Hopper Tensor Cores can apply mixed FP8 and FP16 precisions to dramatically accelerate AI
calculations for transformers.
NVLink Switch System
The NVLink Switch System enables the scaling of multi-GPU input/output (IO) across multiple
servers. The system delivers up to 9X higher bandwidth than InfiniBand HDR on the NVIDIA Ampere
architecture.
NVIDIA Confidential Computing
NVIDIA Confidential Computing is a built-in security feature of Hopper that makes NVIDIA H100 the
world’s first accelerator with confidential computing capabilities. Users can protect the confidentiality
and integrity of their data and applications in use while accessing the unsurpassed acceleration of
H100 GPUs.
Second-Generation Multi-Instance GPU (MIG)
The Hopper architecture’s second-generation MIG supports multi-tenant, multi-user configurations in
virtualized environments, securely partitioning the GPU into isolated, right-size instances to maximize
quality of service (QoS) for 7X more secured tenants.
DPX Instructions
Hopper’s DPX instructions accelerate dynamic programming algorithms by 40X compared to CPUs
and 7X compared to NVIDIA Ampere architecture GPUs. This leads to dramatically faster times in
disease diagnosis, real-time routing optimizations, and graph analytics.
ThinkSystem NVIDIA H100 PCIe Gen5 GPUs 3
The following figure shows the NVIDIA H100 SXM5 4-GPU Board installed in the ThinkSystem SD665-N V3
server
Figure 3. NVIDIA H100 SXM5 4-GPU Board in the ThinkSystem SD665-N V3 server
Technical specifications
The following table lists the GPU processing specifications and performance of the NVIDIA H100 GPU.
ThinkSystem NVIDIA H100 PCIe Gen5 GPUs 4
Table 2. Specifications of the NVIDIA H100 GPU
Feature H100 NVL 94GB PCIe adapter H100 80GB SXM board H100 94GB SXM board
GPU Architecture NVIDIA Hopper NVIDIA Hopper NVIDIA Hopper
Part number 4X67A89325 BQQV or C1HL BUBB
GPUs per part number 1 BQQV: 4
C1HL: 8
4
NVIDIA Tensor Cores 456 fourth-generation Tensor Cores 528 fourth-generation
Tensor Cores
528 fourth-generation
Tensor Cores
NVIDIA CUDA Cores
(shading units)
14,592 FP32 CUDA Cores 16,896 FP32 CUDA
Cores
16,896 FP32 CUDA
Cores
Peak FP64
performance
34 TFLOPS 34 TFLOPS 34 TFLOPS
Peak FP64 Tensor
Core performance
67 TFLOPS 67 TFLOPS 67 TFLOPS
Peak FP32
performance
67 TFLOPS 67 TFLOPS 67 TFLOPS
Peak Tensor Float 32
(TF32) performance
990 TFLOPS* 989 TFLOPS* 989 TFLOPS*
Peak FP16
performance
1,980 TFOPS* 1,979 TFLOPS* 1,979 TFLOPS*
Peak Bfloat16 (BF16)
performance
1,980 TFOPS* 1,979 TFLOPS* 1,979 TFLOPS*
Peak FP8 performance 3,960 TFOPS*
INT8 Integer
Performance
3,960 TOPS* 3,958 TOPS* 3,958 TOPS*
GPU Memory 94 GB HBM3 80GB board (feature
BQQV): 80 GB HBM3
94GB board (feature
BUBB): 90GB HBM2e
Memory Bandwidth 3.9 TB/s 80GB board (feature
BQQV): 3.35 TB/sec
94GB board (feature
BUBB): 2.4 TB/sec
ECC Yes Yes Yes
Interconnect Bandwidth NVLink: 600 GB/sec
PCIe Gen5: 128 GB/sec
NVLink: 900 GB/sec
PCIe Gen5: 128 GB/sec
NVLink: 900 GB/sec
PCIe Gen5: 128 GB/sec
System Interface PCIe Gen 5.0, x16 lanes PCIe Gen 5.0, x16 lanes PCIe Gen 5.0, x16 lanes
Form Factor PCIe full height/length, double width SXM5 SXM5
NVLink support Yes; 3 NVLink Bridge supported per
pair of GPUs (all 3 required)
Yes, integrated Yes, integrated
Multi-Instance GPU
(MIG)
Up to 7 GPU instances, 12GB each Up to 7 GPU instances,
10GB each
Up to 7 GPU instances,
10GB each
Max Power
Consumption
400 W 700 W 700 W
Thermal Solution Passive Water cooled Water cooled
Compute APIs CUDA, DirectCompute, OpenCL,
OpenACC
CUDA, DirectCompute,
OpenCL, OpenACC
CUDA, DirectCompute,
OpenCL, OpenACC
* With structural sparsity enabled
Server support
ThinkSystem NVIDIA H100 PCIe Gen5 GPUs 5
剩余26页未读,继续阅读
资源评论
maming13
- 粉丝: 0
- 资源: 4
上传资源 快速赚钱
- 我的内容管理 展开
- 我的资源 快来上传第一个资源
- 我的收益 登录查看自己的收益
- 我的积分 登录查看自己的积分
- 我的C币 登录后查看C币余额
- 我的收藏
- 我的下载
- 下载帮助
最新资源
- 2025继续教育公需课必修课试题(含答案).pptx
- 2025健康知识竞赛题库(含答案).pptx
- 2025继续教育考试题(含答案).pptx
- 公司中高层管理人员薪酬管理方案.doc
- 公司董事、高级管理人员薪酬考核制度.doc
- 高管人员薪酬方案.doc
- 高管薪酬管理制度.doc
- 高管薪酬设计方案.doc
- 中高层管理人员薪酬管理制度.doc
- 远大公司高层薪酬制度.doc
- 南航高管薪酬管理制度.docx
- 高级管理人员年薪制管理办法.docx
- 委派子公司高管绩效薪酬制度.docx
- 高管人员薪酬与绩效考核管理制度.docx
- 2025交管12123学法减分试题库(含参考答案).pptx
- 2025计算机网络技术考试题(含答案).doc
资源上传下载、课程学习等过程中有任何疑问或建议,欢迎提出宝贵意见哦~我们会及时处理!
点击此处反馈
安全验证
文档复制为VIP权益,开通VIP直接复制
信息提交成功