Dihuni OptiReady Supermicro 9029GP-TNVRT-V16-1 HGX-2 16x NVIDIA Tesla V100 SXM3 32GB NVLink2 GPU 2S Scalable Xeon NVMe 2x10GbE Deep Learning Server

NVSwitch and NVLink Performance

NVSwitch enables every GPU to communicate with every other GPU at full bandwidth of 2.4TB/sec to solve the largest of AI and HPC problems. Every GPU has full access to 0.5TB of aggregate HBM2 memory to handle the most massive of datasets. By enabling a unified server node, NVSwitch dramatically accelerates complex AI and HPC applications.

Designed for Next Generation AI

AI models are exploding in complexity and require large memory, multiple GPUs, and an extremely fast connection between the GPUs to work. With NVSwitch connecting all GPUs and unified memory, HGX-2 provides the power to handle these new models for faster training of advanced AI. A single HGX-2 replaces 300 CPU-powered servers, saving significant cost, space, and energy in the data center.

Ready for Max Density HPC

HPC applications require strong server nodes with the computing power to perform a massive number of calculations per second. Increasing the compute density of each node dramatically reduces the number of servers required, resulting in huge savings in cost, power, and space consumed in the data center. For HPC simulations, high-dimension matrix multiplication requires a processor to fetch data from many neighbors to facilitate computation, making GPUs connected by NVSwitch ideal. A single HGX-2 server replaces 60 CPU-only servers.

X11 Servers Featuring New Intel Skylake Scalable Xeon® Processors

Supermicro’s new X11 servers are engineered to unleash the full performance and rich feature sets on the new Intel® Xeon® Scalable processor family, supporting more cores and higher TDP envelopes of 205 watts and higher, increased number of memory channels and higher bandwidth, more PCI-E 3.0 lanes, 100G/40G/25G/10G Ethernet, 100G EDR InfiniBand (on select servers) and integrated  Intel® Omni-Path Architecture networking fabrics. The elevated compute performance, density, I/O capacity, and efficiency are coupled with industry’s most comprehensive support for NVMe NAND Flash and Intel® Optane SSDs for unprecedented application responsiveness and agility. For exact sever specifications, please see highlights below and also refer to detailed technical specifications.

“Supermicro’s new SuperServer based on the HGX-2 platform will deliver more than double the performance of current systems, which will help enterprises address the rapidly expanding size of AI models that sometimes require weeks to train,” said Charles Liang, president and CEO of Supermicro. “Our new HGX-2 system will enable efficient training of complex models. It combines sixteen Tesla V100 32GB SXM3 GPUs connected via NVLink and NVSwitch to work as a unified 2 PetaFlop accelerator with half a terabyte of aggregate GPU memory to deliver unmatched compute power.”

Charles Liang, President and CEO of Supermicro

Server Systems Management

Supermicro Server Manager (SSM) provides capabilities to monitor the health of server components including memory, hard drives and RAID controllers. It enables the datacenter administrator to monitor and manage power usage across all Supermicro servers allowing users to maximize their CPU payload while mitigating the risk of tripped circuit. Firmware upgrades on Supermicro servers became easier now with a couple of clicks. Administrators can now mount an ISO image on multiple servers and reboot the servers with those images. The tool also provides pre-defined reports and many more features that will make managing Supermicro servers simpler. Download the SSM_brochure for more info or download Supermicro SuperDoctor® device monitoring and management software.

Posted on

Dihuni Introduces Supermicro’s NVIDIA HGX-2 based 16 Tesla V100 32GB SXM2 GPU Server for Deep Learning, AI, HPC and IoT Predictive Analytics

We are pleased to announce our plans to introduce Supermicro’s upcoming NVIDIA® HGX-2 based cloud server platform which the company describes as the world’s most powerful system for artificial intelligence (AI) and high-performance computing (HPC) capable of performing at 2 PetaFLOPS. Continue reading Dihuni Introduces Supermicro’s NVIDIA HGX-2 based 16 Tesla V100 32GB SXM2 GPU Server for Deep Learning, AI, HPC and IoT Predictive Analytics

HPE NVIDIA Tesla V100 GPU 32GB HBM2 Volta CUDA PCIe for Accelerated Machine Deep Learning AI BigData Finance Oil Gas CAD HPC Physics Research

Double Memory than Previous Generation V100

The Tesla V100 GPU, widely adopted by the world’s leading researchers, has received a 2x memory boost to handle the most memory-intensive deep learning and high performance computing workloads.

Now equipped with 32GB of memory, Tesla V100 GPUs will help data scientists train deeper and larger deep learning models that are more accurate than ever. They can also improve the performance of memory-constrained HPC applications by up to 50 percent compared with the previous 16GB version.

GroundBreaking Volta Architecture

By pairing CUDA Cores and Tensor Cores within a unified architecture, a single server with Tesla V100 GPUs can replace hundreds of commodity CPU servers for traditional HPC and Deep Learning. TENSOR CORE Equipped with 640 Tensor Cores, Tesla V100 delivers 125 TeraFLOPS of deep learning performance. That’s 12X Tensor FLOPS for DL Training, and 6X Tensor FLOPS for DL Inference when compared to NVIDIA Pascal™ GPUs.

Next Generation NVLink

NVIDIA NVLink in Tesla V100 delivers 2X higher throughput compared to the previous generation. Up to eight Tesla V100 accelerators can be interconnected at up to 300 GB/s to unleash the highest application performance possible on a single server. HBM2 With a combination of improved raw bandwidth of 900 GB/s and higher DRAM utilization efficiency at 95%, Tesla V100 delivers 1.5X higher memory bandwidth over Pascal GPUs as measured on STREAM.

Maximum Efficiency Mode

The new maximum efficiency mode allows data centers to achieve up to 40% higher compute capacity per rack within the existing power budget. In this mode, Tesla V100 runs at peak processing efficiency, providing up to 80% of the performance at half the power consumption.

Programmability

Tesla V100 is architected from the ground up to simplify programmability. Its new independent thread scheduling enables finer-grain synchronization and improves GPU utilization by sharing resources among small jobs.

CUDA Ready

CUDA® is a parallel computing platform and programming model developed by NVIDIA for general computing on graphical processing units (GPUs). With CUDA, developers are able to dramatically speed up computing applications by harnessing the power of GPUs.

In GPU-accelerated applications, the sequential part of the workload runs on the CPU – which is optimized for single-threaded performance – while the compute intensive portion of the application runs on thousands of GPU cores in parallel. When using CUDA, developers program in popular languages such as C, C++, Fortran, Python and MATLAB and express parallelism through extensions in the form of a few basic keywords.

The CUDA Toolkit from NVIDIA provides everything you need to develop GPU-accelerated applications. The CUDA Toolkit includes GPU-accelerated libraries, a compiler, development tools and the CUDA runtime.

Performance Specifications for NVIDIA Tesla P4, P40 and V100 Accelerators

Tesla V100: The Universal Datacenter GPU Tesla P4 for Ultra-Efficient Scale-Out Servers Tesla P40 for Inference Throughput Servers
Single-Precision Performance (FP32) 14 teraflops (PCIe)
15.7 teraflops (SXM2)
5.5 teraflops 12 teraflops
Half-Precision Performance (FP16) 112 teraflops (PCIe)
125 teraflops (SXM2)
Integer Operations (INT8) 22 TOPS* 47 TOPS*
GPU Memory 16/32 GB HBM2 8 GB 24 GB
Memory Bandwidth 900 GB/s 192 GB/s 346 GB/s
System Interface/Form Factor Dual-Slot, Full-Height PCI Express Form Factor
SXM2 / NVLink
Low-Profile PCI Express Form Factor Dual-Slot, Full-Height PCI Express Form Factor
Power 250W (PCIe)
300W (SXM2)
50 W/75 W 250 W
Hardware-Accelerated Video Engine 1x Decode Engine, 2x Encode Engines 1x Decode Engine, 2x Encode Engines

*Tera-Operations per Second with Boost Clock Enabled

HPE NVIDIA Tesla V100 GPU 16GB HBM2 Volta CUDA PCIe for Accelerated Machine Deep Learning AI BigData Finance Oil Gas CAD HPC Physics Research

GroundBreaking Volta Architecture

By pairing CUDA Cores and Tensor Cores within a unified architecture, a single server with Tesla V100 GPUs can replace hundreds of commodity CPU servers for traditional HPC and Deep Learning. TENSOR CORE Equipped with 640 Tensor Cores, Tesla V100 delivers 125 TeraFLOPS of deep learning performance. That’s 12X Tensor FLOPS for DL Training, and 6X Tensor FLOPS for DL Inference when compared to NVIDIA Pascal™ GPUs.

Next Generation NVLink

NVIDIA NVLink in Tesla V100 delivers 2X higher throughput compared to the previous generation. Up to eight Tesla V100 accelerators can be interconnected at up to 300 GB/s to unleash the highest application performance possible on a single server. HBM2 With a combination of improved raw bandwidth of 900 GB/s and higher DRAM utilization efficiency at 95%, Tesla V100 delivers 1.5X higher memory bandwidth over Pascal GPUs as measured on STREAM.

Maximum Efficiency Mode

The new maximum efficiency mode allows data centers to achieve up to 40% higher compute capacity per rack within the existing power budget. In this mode, Tesla V100 runs at peak processing efficiency, providing up to 80% of the performance at half the power consumption.

Programmability

Tesla V100 is architected from the ground up to simplify programmability. Its new independent thread scheduling enables finer-grain synchronization and improves GPU utilization by sharing resources among small jobs.

CUDA Ready

CUDA® is a parallel computing platform and programming model developed by NVIDIA for general computing on graphical processing units (GPUs). With CUDA, developers are able to dramatically speed up computing applications by harnessing the power of GPUs.

In GPU-accelerated applications, the sequential part of the workload runs on the CPU – which is optimized for single-threaded performance – while the compute intensive portion of the application runs on thousands of GPU cores in parallel. When using CUDA, developers program in popular languages such as C, C++, Fortran, Python and MATLAB and express parallelism through extensions in the form of a few basic keywords.

The CUDA Toolkit from NVIDIA provides everything you need to develop GPU-accelerated applications. The CUDA Toolkit includes GPU-accelerated libraries, a compiler, development tools and the CUDA runtime.

Performance Specifications for NVIDIA Tesla P4, P40 and V100 Accelerators

Tesla V100: The Universal Datacenter GPU Tesla P4 for Ultra-Efficient Scale-Out Servers Tesla P40 for Inference Throughput Servers
Single-Precision Performance (FP32) 14 teraflops (PCIe)
15.7 teraflops (SXM2)
5.5 teraflops 12 teraflops
Half-Precision Performance (FP16) 112 teraflops (PCIe)
125 teraflops (SXM2)
Integer Operations (INT8) 22 TOPS* 47 TOPS*
GPU Memory 16/32 GB HBM2 8 GB 24 GB
Memory Bandwidth 900 GB/s 192 GB/s 346 GB/s
System Interface/Form Factor Dual-Slot, Full-Height PCI Express Form Factor
SXM2 / NVLink
Low-Profile PCI Express Form Factor Dual-Slot, Full-Height PCI Express Form Factor
Power 250W (PCIe)
300W (SXM2)
50 W/75 W 250 W
Hardware-Accelerated Video Engine 1x Decode Engine, 2x Encode Engines 1x Decode Engine, 2x Encode Engines

*Tera-Operations per Second with Boost Clock Enabled

Lenovo ThinkSystem NVIDIA Tesla V100 GPU 16GB HBM2 Volta CUDA PCIe for Accelerated Machine Deep Learning AI Finance Oil Gas CAD HPC Physics Research

Lenovo ThinkSystem servers support GPU technology to accelerate different computing workloads, maximize performance for graphic design, virtualization, artificial intelligence and high performance computing applications in Lenovo servers.

The following table summarizes the server support for the GPUs. The numbers listed in the server columns represent the number of GPUs supported.

GroundBreaking Volta Architecture

By pairing CUDA Cores and Tensor Cores within a unified architecture, a single server with Tesla V100 GPUs can replace hundreds of commodity CPU servers for traditional HPC and Deep Learning. TENSOR CORE Equipped with 640 Tensor Cores, Tesla V100 delivers 125 TeraFLOPS of deep learning performance. That’s 12X Tensor FLOPS for DL Training, and 6X Tensor FLOPS for DL Inference when compared to NVIDIA Pascal™ GPUs.

Next Generation NVLink

NVIDIA NVLink in Tesla V100 delivers 2X higher throughput compared to the previous generation. Up to eight Tesla V100 accelerators can be interconnected at up to 300 GB/s to unleash the highest application performance possible on a single server. HBM2 With a combination of improved raw bandwidth of 900 GB/s and higher DRAM utilization efficiency at 95%, Tesla V100 delivers 1.5X higher memory bandwidth over Pascal GPUs as measured on STREAM.

Maximum Efficiency Mode

The new maximum efficiency mode allows data centers to achieve up to 40% higher compute capacity per rack within the existing power budget. In this mode, Tesla V100 runs at peak processing efficiency, providing up to 80% of the performance at half the power consumption.

Programmability

Tesla V100 is architected from the ground up to simplify programmability. Its new independent thread scheduling enables finer-grain synchronization and improves GPU utilization by sharing resources among small jobs.

CUDA Ready

CUDA® is a parallel computing platform and programming model developed by NVIDIA for general computing on graphical processing units (GPUs). With CUDA, developers are able to dramatically speed up computing applications by harnessing the power of GPUs.

In GPU-accelerated applications, the sequential part of the workload runs on the CPU – which is optimized for single-threaded performance – while the compute intensive portion of the application runs on thousands of GPU cores in parallel. When using CUDA, developers program in popular languages such as C, C++, Fortran, Python and MATLAB and express parallelism through extensions in the form of a few basic keywords.

The CUDA Toolkit from NVIDIA provides everything you need to develop GPU-accelerated applications. The CUDA Toolkit includes GPU-accelerated libraries, a compiler, development tools and the CUDA runtime.

Performance Specifications for NVIDIA Tesla P4, P40 and V100 Accelerators

Tesla V100: The Universal Datacenter GPU Tesla P4 for Ultra-Efficient Scale-Out Servers Tesla P40 for Inference Throughput Servers
Single-Precision Performance (FP32) 14 teraflops (PCIe)
15.7 teraflops (SXM2)
5.5 teraflops 12 teraflops
Half-Precision Performance (FP16) 112 teraflops (PCIe)
125 teraflops (SXM2)
Integer Operations (INT8) 22 TOPS* 47 TOPS*
GPU Memory 16/32 GB HBM2 8 GB 24 GB
Memory Bandwidth 900 GB/s 192 GB/s 346 GB/s
System Interface/Form Factor Dual-Slot, Full-Height PCI Express Form Factor
SXM2 / NVLink
Low-Profile PCI Express Form Factor Dual-Slot, Full-Height PCI Express Form Factor
Power 250W (PCIe)
300W (SXM2)
50 W/75 W 250 W
Hardware-Accelerated Video Engine 1x Decode Engine, 2x Encode Engines 1x Decode Engine, 2x Encode Engines

*Tera-Operations per Second with Boost Clock Enabled

Lenovo ThinkSystem NVIDIA Tesla V100 GPU 32GB HBM2 Volta CUDA PCIe for Accelerated Machine Deep Learning AI Finance Oil Gas CAD HPC Physics Research

Lenovo ThinkSystem servers support GPU technology to accelerate different computing workloads, maximize performance for graphic design, virtualization, artificial intelligence and high performance computing applications in Lenovo servers.

The following table summarizes the server support for the GPUs. The numbers listed in the server columns represent the number of GPUs supported.

Double Memory than Previous Generation V100

The Tesla V100 GPU, widely adopted by the world’s leading researchers, has received a 2x memory boost to handle the most memory-intensive deep learning and high performance computing workloads.

Now equipped with 32GB of memory, Tesla V100 GPUs will help data scientists train deeper and larger deep learning models that are more accurate than ever. They can also improve the performance of memory-constrained HPC applications by up to 50 percent compared with the previous 16GB version.

GroundBreaking Volta Architecture

By pairing CUDA Cores and Tensor Cores within a unified architecture, a single server with Tesla V100 GPUs can replace hundreds of commodity CPU servers for traditional HPC and Deep Learning. TENSOR CORE Equipped with 640 Tensor Cores, Tesla V100 delivers 125 TeraFLOPS of deep learning performance. That’s 12X Tensor FLOPS for DL Training, and 6X Tensor FLOPS for DL Inference when compared to NVIDIA Pascal™ GPUs.

Next Generation NVLink

NVIDIA NVLink in Tesla V100 delivers 2X higher throughput compared to the previous generation. Up to eight Tesla V100 accelerators can be interconnected at up to 300 GB/s to unleash the highest application performance possible on a single server. HBM2 With a combination of improved raw bandwidth of 900 GB/s and higher DRAM utilization efficiency at 95%, Tesla V100 delivers 1.5X higher memory bandwidth over Pascal GPUs as measured on STREAM.

Maximum Efficiency Mode

The new maximum efficiency mode allows data centers to achieve up to 40% higher compute capacity per rack within the existing power budget. In this mode, Tesla V100 runs at peak processing efficiency, providing up to 80% of the performance at half the power consumption.

Programmability

Tesla V100 is architected from the ground up to simplify programmability. Its new independent thread scheduling enables finer-grain synchronization and improves GPU utilization by sharing resources among small jobs.

CUDA Ready

CUDA® is a parallel computing platform and programming model developed by NVIDIA for general computing on graphical processing units (GPUs). With CUDA, developers are able to dramatically speed up computing applications by harnessing the power of GPUs.

In GPU-accelerated applications, the sequential part of the workload runs on the CPU – which is optimized for single-threaded performance – while the compute intensive portion of the application runs on thousands of GPU cores in parallel. When using CUDA, developers program in popular languages such as C, C++, Fortran, Python and MATLAB and express parallelism through extensions in the form of a few basic keywords.

The CUDA Toolkit from NVIDIA provides everything you need to develop GPU-accelerated applications. The CUDA Toolkit includes GPU-accelerated libraries, a compiler, development tools and the CUDA runtime.

Performance Specifications for NVIDIA Tesla P4, P40 and V100 Accelerators

Tesla V100: The Universal Datacenter GPU Tesla P4 for Ultra-Efficient Scale-Out Servers Tesla P40 for Inference Throughput Servers
Single-Precision Performance (FP32) 14 teraflops (PCIe)
15.7 teraflops (SXM2)
5.5 teraflops 12 teraflops
Half-Precision Performance (FP16) 112 teraflops (PCIe)
125 teraflops (SXM2)
Integer Operations (INT8) 22 TOPS* 47 TOPS*
GPU Memory 16/32 GB HBM2 8 GB 24 GB
Memory Bandwidth 900 GB/s 192 GB/s 346 GB/s
System Interface/Form Factor Dual-Slot, Full-Height PCI Express Form Factor
SXM2 / NVLink
Low-Profile PCI Express Form Factor Dual-Slot, Full-Height PCI Express Form Factor
Power 250W (PCIe)
300W (SXM2)
50 W/75 W 250 W
Hardware-Accelerated Video Engine 1x Decode Engine, 2x Encode Engines 1x Decode Engine, 2x Encode Engines

*Tera-Operations per Second with Boost Clock Enabled

Dihuni OptiReady Supermicro 4029GP-TRT2-V100-1 4U 5x NVIDIA Tesla V100 32GB GPU 2S Xeon 4116 2.1GHz 128GB 250GBSSD 1TBHDD 2x10GbE Deep Learning Server

X11 Servers Featuring New Intel Skylake Scalable Xeon® Processors

Supermicro’s new X11 servers are engineered to unleash the full performance and rich feature sets on the new Intel® Xeon® Scalable processor family, supporting more cores and higher TDP envelopes of 205 watts and higher, increased number of memory channels and higher bandwidth, more PCI-E 3.0 lanes, 100G/40G/25G/10G Ethernet, 100G EDR InfiniBand (on select servers) and integrated  Intel® Omni-Path Architecture networking fabrics. The elevated compute performance, density, I/O capacity, and efficiency are coupled with industry’s most comprehensive support for NVMe NAND Flash and Intel® Optane SSDs for unprecedented application responsiveness and agility. For exact sever specifications, please see highlights below and also refer to detailed technical specifications.

“At Supermicro, we understand that customers need the newest technologies as early as possible to drive leading performance and improved TCO. With the industry’s strongest and broadest product line, our designs not only take full advantage of Xeon Scalable Processors’ new features such as three UPI, faster DIMMs and more core count per socket, but they also fully support NVMe through unique non-blocking architectures to achieve the best data bandwidth and IOPS.  For instance, one Supermicro 2U storage server can deliver over 16 million IOPS!”

“Supermicro designs the most application-optimized GPU systems and offers the widest selection of GPU-optimized servers and workstations in the industry. Our high performance computing solutions enable deep learning, engineering and scientific fields to scale out their compute clusters to accelerate their most demanding workloads and achieve fastest time-to-results with maximum performance per watt, per square foot and per dollar. With our latest innovations incorporating the new NVIDIA V100 PCI-E and V100 SXM2 GPUs in performance-optimized 1U and 4U systems with next-generation NVLink, our customers can accelerate their applications and innovations to help solve the world’s most complex and challenging problems.”  

Charles Liang, President and CEO of Supermicro

Support for 8 Double Width GPUs for Deep Learning

The 4029GP-TRT2 takes full advantage of the new Xeon Scalable Processor Family  PCIe lanes to support 8 double-width GPUs to deliver a very high performance Artificial Intelligence and Deep Learning system suitable for autonomous cars, molecular dynamics, computational biology, fluid simulation, advanced physics and Internet of Things (IoT) and Big Data Analytics etc. With NVIDIA Tesla cards, this server delivers unparalleled acceleration for compute intensive applications.

Server Systems Management

Supermicro Server Manager (SSM) provides capabilities to monitor the health of server components including memory, hard drives and RAID controllers. It enables the datacenter administrator to monitor and manage power usage across all Supermicro servers allowing users to maximize their CPU payload while mitigating the risk of tripped circuit. Firmware upgrades on Supermicro servers became easier now with a couple of clicks. Administrators can now mount an ISO image on multiple servers and reboot the servers with those images. The tool also provides pre-defined reports and many more features that will make managing Supermicro servers simpler. Download the SSM_brochure for more info or download Supermicro SuperDoctor® device monitoring and management software.

Technical Specifications

Mfr Part # SYS-4029GP-TRT2
Motherboard Super X11DPG-OT-CPU
CPU Dual Socket P (LGA 3647); Intel® Xeon® Scalable Processors,
Dual UPI up to 10.4GT/s; Dual UPI up to 10.4GT/s; Support CPU TDP 70-205W2 x Intel Skylake Xeon Silver 4116 2.1 GHz 12 Core CPU Installed
Cores Up to 28 Cores with Intel® HT Technology
GPU / Coprocessor Support Please refer to: Compatible GPU list
Memory Capacity 24 DIMM slots; Up to 3TB ECC 3DS LRDIMM, 1TB ECC RDIMM, DDR4 up to 2666MHz

128 GB DDR4-2666MHz (32GB x 4) Installed

Memory Type 2666/2400/2133MHz ECC DDR4 SDRAM
Chipset Intel® C622 chipset
SATA SATA3 (6Gbps) with RAID 0, 1, 5, 10
Network Controllers Dual Port 10GbE from C622
IPMI Support for Intelligent Platform Management Interface v.2.0; IPMI 2.0 with virtual media over LAN and KVM-over-LAN support
Graphics ASPEED AST2500 BMC
SATA 10 SATA3 (6Gbps) ports
LAN 2 RJ45 10GBase-T LAN ports; 1 RJ45 Dedicated IPMI LAN port
USB 4 USB 3.0 ports (rear)
Video 1 VGA Connector
COM Port 1 COM port (rear)
BIOS Type AMI 32Mb SPI Flash ROM
Software Intel® Node Manager; IPMI 2.0; KVM with dedicated LAN; SSM, SPM, SUM; ,; SuperDoctor® 5; Watchdog
CPU Monitors for CPU Cores, Chipset Voltages, Memory.; 4+1 Phase-switching voltage regulator
FAN Fans with tachometer monitoring; Status monitor for speed control; Pulse Width Modulated (PWM) fan connectors
Temperature Monitoring for CPU and chassis environment; Thermal Control for fan connectors
Form Factor 4U Rackmountable; Rackmount Kit (MCP-290-00057-0N)
Model CSE-418GTS-R4000B
Height 7.0″ (178mm)
Width 17.2″ (437mm)
Depth 29″ (737mm)
Net Weight: 80 lbs (36.2 kg); Gross Weight: 135 lbs (61.2 kg)
Available Colors Black
Hot-swap Up to 24 Hot-swap 2.5″ SAS/SATA drive bays; 8x 2.5″ drives supported natively

  • 1 x SamsungSM863a 240GB SATA 6Gb/s,VNAND,V48,2.5″,7mm SSD Installed
  • 1 x Seagate 2.5″ 1TB SATA 6Gb/s, 7.2K RPM, 4kN, 128MB HDD Installed
PCI-Express 11 PCI-E 3.0 x16 (FH, FL) slots; 1 PCI-E 3.0 x8 (FH, FL, in x16) slot

  • 5 x NVIDIA V100 32GB PCIe3.0 GPU Installed
Fans 8 Hot-swap 92mm cooling fans
Shrouds 1 Air Shroud (MCP-310-41808-0B)
Total Output Power 1000W/1800W/1980W/2000W
Dimension
(W x H x L)
73.5 x 40 x 265 mm
Input 100-120Vac / 12.5-9.5A / 50-60Hz; 200-220Vac / 10-9.5A / 50-60Hz; 220-230Vac / 10-9.8A / 50-60Hz; 230-240Vac / 10-9.8A / 50-60Hz; 200-240Vac / 11.8-9.8A / 50-60Hz (UL/cUL only)
+12V Max: 83.3A / Min: 0A (100-120Vac); Max: 150A / Min: 0A (200-220Vac); Max: 165A / Min: 0A (220-230Vac); Max: 166.7A / Min: 0A (230-240Vac); Max: 166.7A / Min: 0A (200-240Vac) (UL/cUL only)
12Vsb Max: 2.1A / Min: 0A
Output Type 25 Pairs Gold Finger Connector
Certification Titanium Level; [ Test Report ]
RoHS RoHS Compliant
Environmental Spec. Operating Temperature:
10°C ~ 35°C (50°F ~ 95°F); Non-operating Temperature:
-40°C to 60°C (-40°F to 140°F); Operating Relative Humidity:
8% to 90% (non-condensing); Non-operating Relative Humidity:
5% to 95% (non-condensing)

Dihuni OptiReady Supermicro 4029GP-TRT2-V100-1 4U 10 NVIDIA Tesla V100 32GB GPU 2S Xeon 4116 2.1GHz 256GB 250GBSSD 1TBHDD 2x10GbE Deep Learning Server

X11 Servers Featuring New Intel Skylake Scalable Xeon® Processors

Supermicro’s new X11 servers are engineered to unleash the full performance and rich feature sets on the new Intel® Xeon® Scalable processor family, supporting more cores and higher TDP envelopes of 205 watts and higher, increased number of memory channels and higher bandwidth, more PCI-E 3.0 lanes, 100G/40G/25G/10G Ethernet, 100G EDR InfiniBand (on select servers) and integrated  Intel® Omni-Path Architecture networking fabrics. The elevated compute performance, density, I/O capacity, and efficiency are coupled with industry’s most comprehensive support for NVMe NAND Flash and Intel® Optane SSDs for unprecedented application responsiveness and agility. For exact sever specifications, please see highlights below and also refer to detailed technical specifications.

“At Supermicro, we understand that customers need the newest technologies as early as possible to drive leading performance and improved TCO. With the industry’s strongest and broadest product line, our designs not only take full advantage of Xeon Scalable Processors’ new features such as three UPI, faster DIMMs and more core count per socket, but they also fully support NVMe through unique non-blocking architectures to achieve the best data bandwidth and IOPS.  For instance, one Supermicro 2U storage server can deliver over 16 million IOPS!”

“Supermicro designs the most application-optimized GPU systems and offers the widest selection of GPU-optimized servers and workstations in the industry. Our high performance computing solutions enable deep learning, engineering and scientific fields to scale out their compute clusters to accelerate their most demanding workloads and achieve fastest time-to-results with maximum performance per watt, per square foot and per dollar. With our latest innovations incorporating the new NVIDIA V100 PCI-E and V100 SXM2 GPUs in performance-optimized 1U and 4U systems with next-generation NVLink, our customers can accelerate their applications and innovations to help solve the world’s most complex and challenging problems.”  

Charles Liang, President and CEO of Supermicro

Support for 8 Double Width GPUs for Deep Learning

The 4029GP-TRT2 takes full advantage of the new Xeon Scalable Processor Family  PCIe lanes to support 8 double-width GPUs to deliver a very high performance Artificial Intelligence and Deep Learning system suitable for autonomous cars, molecular dynamics, computational biology, fluid simulation, advanced physics and Internet of Things (IoT) and Big Data Analytics etc. With NVIDIA Tesla cards, this server delivers unparalleled acceleration for compute intensive applications.

Server Systems Management

Supermicro Server Manager (SSM) provides capabilities to monitor the health of server components including memory, hard drives and RAID controllers. It enables the datacenter administrator to monitor and manage power usage across all Supermicro servers allowing users to maximize their CPU payload while mitigating the risk of tripped circuit. Firmware upgrades on Supermicro servers became easier now with a couple of clicks. Administrators can now mount an ISO image on multiple servers and reboot the servers with those images. The tool also provides pre-defined reports and many more features that will make managing Supermicro servers simpler. Download the SSM_brochure for more info or download Supermicro SuperDoctor® device monitoring and management software.

Technical Specifications

Mfr Part # SYS-4029GP-TRT2
Motherboard Super X11DPG-OT-CPU
CPU Dual Socket P (LGA 3647); Intel® Xeon® Scalable Processors,
Dual UPI up to 10.4GT/s; Dual UPI up to 10.4GT/s; Support CPU TDP 70-205W2 x Intel Skylake Xeon Silver 4116 2.1 GHz 12 Core CPU Installed
Cores Up to 28 Cores with Intel® HT Technology
GPU / Coprocessor Support Please refer to: Compatible GPU list
Memory Capacity 24 DIMM slots; Up to 3TB ECC 3DS LRDIMM, 1TB ECC RDIMM, DDR4 up to 2666MHz

256 GB DDR4-2666MHz (32GB x 8) Installed

Memory Type 2666/2400/2133MHz ECC DDR4 SDRAM
Chipset Intel® C622 chipset
SATA SATA3 (6Gbps) with RAID 0, 1, 5, 10
Network Controllers Dual Port 10GbE from C622
IPMI Support for Intelligent Platform Management Interface v.2.0; IPMI 2.0 with virtual media over LAN and KVM-over-LAN support
Graphics ASPEED AST2500 BMC
SATA 10 SATA3 (6Gbps) ports
LAN 2 RJ45 10GBase-T LAN ports; 1 RJ45 Dedicated IPMI LAN port
USB 4 USB 3.0 ports (rear)
Video 1 VGA Connector
COM Port 1 COM port (rear)
BIOS Type AMI 32Mb SPI Flash ROM
Software Intel® Node Manager; IPMI 2.0; KVM with dedicated LAN; SSM, SPM, SUM; ,; SuperDoctor® 5; Watchdog
CPU Monitors for CPU Cores, Chipset Voltages, Memory.; 4+1 Phase-switching voltage regulator
FAN Fans with tachometer monitoring; Status monitor for speed control; Pulse Width Modulated (PWM) fan connectors
Temperature Monitoring for CPU and chassis environment; Thermal Control for fan connectors
Form Factor 4U Rackmountable; Rackmount Kit (MCP-290-00057-0N)
Model CSE-418GTS-R4000B
Height 7.0″ (178mm)
Width 17.2″ (437mm)
Depth 29″ (737mm)
Net Weight: 80 lbs (36.2 kg); Gross Weight: 135 lbs (61.2 kg)
Available Colors Black
Hot-swap Up to 24 Hot-swap 2.5″ SAS/SATA drive bays; 8x 2.5″ drives supported natively

  • 1 x SamsungSM863a 240GB SATA 6Gb/s,VNAND,V48,2.5″,7mm SSD Installed
  • 1 x Seagate 2.5″ 1TB SATA 6Gb/s, 7.2K RPM, 4kN, 128MB HDD Installed
PCI-Express 11 PCI-E 3.0 x16 (FH, FL) slots; 1 PCI-E 3.0 x8 (FH, FL, in x16) slot

  • 10 x NVIDIA V100 32GB PCIe3.0 GPU Installed
Fans 8 Hot-swap 92mm cooling fans
Shrouds 1 Air Shroud (MCP-310-41808-0B)
Total Output Power 1000W/1800W/1980W/2000W
Dimension
(W x H x L)
73.5 x 40 x 265 mm
Input 100-120Vac / 12.5-9.5A / 50-60Hz; 200-220Vac / 10-9.5A / 50-60Hz; 220-230Vac / 10-9.8A / 50-60Hz; 230-240Vac / 10-9.8A / 50-60Hz; 200-240Vac / 11.8-9.8A / 50-60Hz (UL/cUL only)
+12V Max: 83.3A / Min: 0A (100-120Vac); Max: 150A / Min: 0A (200-220Vac); Max: 165A / Min: 0A (220-230Vac); Max: 166.7A / Min: 0A (230-240Vac); Max: 166.7A / Min: 0A (200-240Vac) (UL/cUL only)
12Vsb Max: 2.1A / Min: 0A
Output Type 25 Pairs Gold Finger Connector
Certification Titanium Level; [ Test Report ]
RoHS RoHS Compliant
Environmental Spec. Operating Temperature:
10°C ~ 35°C (50°F ~ 95°F); Non-operating Temperature:
-40°C to 60°C (-40°F to 140°F); Operating Relative Humidity:
8% to 90% (non-condensing); Non-operating Relative Humidity:
5% to 95% (non-condensing)

Dihuni OptiReady Supermicro 7049GP-TRT-V100-1 2 x NVIDIA Tesla V100 32GB GPU 4U Tower 2S Xeon 4114 96GB 480GB SSD 4TB HDD 2x10GbE Deep Learning Server

X11 Servers Featuring New Intel Skylake Scalable Xeon® Processors

Supermicro’s new X11 servers are engineered to unleash the full performance and rich feature sets on the new Intel® Xeon® Scalable processor family, supporting more cores and higher TDP envelopes of 205 watts and higher, increased number of memory channels and higher bandwidth, more PCI-E 3.0 lanes, 100G/40G/25G/10G Ethernet, 100G EDR InfiniBand (on select servers) and integrated  Intel® Omni-Path Architecture networking fabrics. The elevated compute performance, density, I/O capacity, and efficiency are coupled with industry’s most comprehensive support for NVMe NAND Flash and Intel® Optane SSDs for unprecedented application responsiveness and agility. For exact sever specifications, please see highlights below and also refer to detailed technical specifications.

“At Supermicro, we understand that customers need the newest technologies as early as possible to drive leading performance and improved TCO. With the industry’s strongest and broadest product line, our designs not only take full advantage of Xeon Scalable Processors’ new features such as three UPI, faster DIMMs and more core count per socket, but they also fully support NVMe through unique non-blocking architectures to achieve the best data bandwidth and IOPS.  For instance, one Supermicro 2U storage server can deliver over 16 million IOPS!”

“We are excited to preview our X11 Ultra, TwinPro™, BigTwin™, SuperBlade® and many more new designs based on the new Intel® Xeon® Processor Scalable Family processors with CPUs that can provide up to 3.9x higher virtualized throughput,”  

Charles Liang, President and CEO of Supermicro

Server Systems Management

Supermicro Server Manager (SSM) provides capabilities to monitor the health of server components including memory, hard drives and RAID controllers. It enables the datacenter administrator to monitor and manage power usage across all Supermicro servers allowing users to maximize their CPU payload while mitigating the risk of tripped circuit. Firmware upgrades on Supermicro servers became easier now with a couple of clicks. Administrators can now mount an ISO image on multiple servers and reboot the servers with those images. The tool also provides pre-defined reports and many more features that will make managing Supermicro servers simpler. Download the SSM_brochure for more info or download Supermicro SuperDoctor® device monitoring and management software.

Technical Specifications

Mfr Part # SYS-7049GP-TRT (Black)
CPU Dual Socket P (LGA 3647); Intel® Xeon® Scalable Processors,
3 UPI up to 10.4GT/s; Support CPU TDP 70-205W with IVR 2 x Intel Xeon Skylake Gold 4114 10 Core 2.2 GHz CPU Installed
Cores Up to 28 Cores with Intel® HT Technology
GPU GPU Support Matrix
Memory Capacity 16 DIMM slots; Up to 2TB ECC 3DS LRDIMM, 1TB ECC RDIMM, DDR4 up to 2666MHz

96GB ECC DDR4-2666MHz installed (8GB x 12); expandable up to 2TB

Memory Type 2666/2400/2133MHz ECC DDR4 SDRAM
Chipset Intel® C621 chipset
SATA SATA3 (6Gbps); RAID 0, 1, 5, 10
Network Controllers Intel® X550 Dual Port 10GBase-T; Virtual Machine Device Queues reduce I/O overhead; Supports 10GBASE-T, 100BASE-TX, and 1000BASE-T, RJ45 output
IPMI Support for Intelligent Platform Management Interface v.2.0; IPMI 2.0 with virtual media over LAN and KVM-over-LAN support
Graphics ASPEED AST2500 BMC
SATA 10 SATA3 (6Gbps) ports
LAN 2 RJ45 10GBase-T ports; 1 RJ45 Dedicated IPMI LAN port
USB 5 USB 3.0 ports (2 rear, 2 via header, 1 Type A); 4 USB 2.0 ports (2 rear, 2 via headers)
Video 1 VGA port
Serial Port/Header 2 COM ports (1 rear, 1 Header)
BIOS Type AMI 32Mb SPI Flash ROM
Software Intel® Node Manager; IPMI 2.0; SSM, SPM, SUM; SuperDoctor® 5
CPU Monitors for CPU Cores, Chipset Voltages, Memory.; 4+1 Phase-switching voltage regulator
FAN Fans with tachometer monitoring; Status monitor for speed control; Pulse Width Modulated (PWM) fan connectors
Temperature Monitoring for CPU and chassis environment; Thermal Control for 8x fan connectors
Form Factor 4U Rackmountable / Tower; Optional Rackmount Kit
Model CSE-747BTS-R2K20BP
Height 18.2″ (462mm)
Width 7.0″ (178mm)
Depth 26.5″ (673mm)
Package 27″ (H) x 13″ (W) x 38″ (D)
Weight Net Weight: 46 lbs (20.9 kg); Gross Weight: 62 lbs (28.1 kg)
Available Colors Dark Gray
Buttons Power On/Off button; System Reset button
LEDs Power status LED; Hard drive activity LED; Network activity LEDs; System Overheat & Power Fail LED
Ports 2 Front USB 3.0 Ports
PCI-Express 6 PCI-E 3.0 x16 (double-width) slots; 1 PCI-E 3.0 x4 (in x8 slot)

2 x NVIDIA Tesla V100 32GB PCI-E GPU Installed

Hot-swap 8 Hot-swap 3.5″ drive bays

  • 1 x Seagate 2.5″ 480GB SATA 6Gb/s, 7.0mm, 16nm,0.6 DWPD SSD Installed
  • 1 x Toshiba 3.5″ 4TB SATA 6Gb/s 7.2K RPM 128M 512E (Tomcat) HDD Installed
Fans 4 Heavy duty fans; 4 Rear exhaust fans
Heatsink 2 Active heatsink with optimal fan speed control
Total Output Power and Input 1200W with Input 100-127Vac; 1800W with Input 200-220Vac; 1980W with Input 220-230Vac; 2090W with Input 230-240Vac; 2090W with Input 200-220Vac; 2200W with Input 220-240Vac (for UL/cUL use only); 2090W with Input 230-240Vdc (for CCC only)
AC Input Frequency 50-60Hz
Dimension
(W x H x L)
76 x 40 x 336 mm
+12V Max: 100A / Min: 0A (100-127Vac); Max: 150A / Min: 0A (200-220Vac); Max: 165A / Min: 0A (220-230Vac); Max: 174.17A / Min: 0A (230-240Vac); Max: 183.3A / Min: 0A (220-240Vac)
5VSB Max: 1A / Min: 0A
Output Type Backplanes (gold finger)
Certification UL/cUL/CB/BSMI/CE/CCC; Titanium Level
RoHS RoHS Compliant
Environmental Spec. Operating Temperature:
10°C ~ 35°C (50°F ~ 95°F); Non-operating Temperature:
-40°C to 60°C (-40°F to 140°F); Operating Relative Humidity:
8% to 90% (non-condensing); Non-operating Relative Humidity:
5% to 95% (non-condensing)

Tyan B8021G88V2HR-2T-N AMD EPYC Max 6 x NVIDIA Tesla V100 P100 P40 P4 and GTX 1080Ti GPU 2TB ECC 7xPCIe 2×2.5SATA 2xNVMe 2x10GbE Deep Learning Server

TYAN GPU Servers

TYAN offers a wide range of GPU (graphics processing unit) computing platforms that are designed for High Performance Computing (HPC) and massive parallel computing environments. TYAN’s GA88B8021 (B8021G88V2HR-2T-N) GPU platform supports up to 6 GPUs inside traditional 1U rackmount chassis, which helps reduce the rack space requirements, required for high performance applications. With the right workload, these platforms can offer higher performance and reduce power consumption when compared to traditional CPU centric platforms.

Support for NVIDIA® Tesla® GPU Accelerators:

TYAN and NVIDIA have joined together to deliver a GPU computing platform that provides double-precision performance that is both cost-effective and energy-efficient. With support for NVIDIA Tesla GPUs, TYAN’s GPU Computing platforms offer massively parallel compute power to solve the most computationally-intensive challenges.

Target applications for TYAN GPU platforms:

  • High Performance GPU frameworks
  • Cryptography
  • Oil & Gas Discovery
  • Particle Physics
  • Genetic Sequencing
  • Cryptocurrency Mining
  • Large Scale Facial Recognition
  • Machine Learning and Artificial Intelligence

Featuring New AMD EPYC™ 7000 Series Processors

Tyan’s new AMD servers are engineered to unleash the full performance and rich feature sets on the new AMD EPYC™ 7000 Series processor family with more cores, more memory, more I/O, and more security. AMD EPYC processors are based on the 14nm “Zen” x86 core architecture with 32 cores and 64 threads, featuring 8 memory channels and up to16 DIMMs per socket. The new powerful processor allows Tyan to offer the optimal balance of computing, memory and I/O for the customers. By adopting the latest AMD EPYC processor technology, Tyan brings high memory channel bandwidth and PCI Express high speed I/O connectivity features to enterprises and data centers. With servers built on EPYC technology, cloud environments can drive greater scale and performance, virtualized data centers can further increase consolidation ratios while delivering better performing virtual machines and Big Data, and analytics environments can collect and analyze larger data sets much faster. High performance applications in research labs can solve complex problem sets in a significantly accelerated manner. EPYC, with all the critical compute, memory, I/O, and security resources brought together in the SoC with the right ratios, delivers industry-leading performance and enables lower TCO. With the flexibility to choose from 8 to 32 cores, EPYC enables you to deploy the right hardware platforms to meet your workload needs from virtualized infrastructure to large-scale big-data and analytics platforms and legacy line-of-business applications. For exact sever specifications, please see highlights below and also refer to detailed technical specifications.

Tyan’s Transport HX GA88-B8021 offers outstanding computational power for a single-socket server, with support for 8, 16, 24, or 32-Core processors with up to 64 threads. The memory capacity of the server is also incredibly high for a single socket server, able to support up to 2TB of RAM with every processor in the AMD EPYC SKU stack.

Technical Specifications

TYAN GA88B8021 (B8021G88V2HR-2T-N) Specifications
System Form Factor 1U Rackmount
Chassis Model GA88
Dimension (D x W x H) 34.84″ x 17.24″ x 1.71″ (885 x 438 x 43.5mm)
Motherboard Name S8021GM2NR-2T
Gross Weight 26.5 kg (58.5 lbs)
Net weight 13.5 kg (30 lbs)
Front Panel Buttons (1) RST / (1) PWR w/ LED / (1) UID
LEDs (1) PWR / (2) LAN / (1) UID / (1) BMC event
I/O Ports (2) USB 3.0 ports
External Drive Bay Q’ty / Type (2) 2.5″ Hot-Swap SSD/HDD
HDD Backplane Support SAS 12Gb/s /SATA 6Gb/s
Supported HDD Interface (2) SAS 12Gb/s / SATA 6Gb/s
System Cooling Configuration Redundancy Yes
FAN (10) 4cm fans
Heat Sink (1) Passive CPU Heat sink
Power Supply Type ERP1U
Input Range AC 100-127V/12A / AC 200-240V/9.48A
Frequency 47-63 Hz
Output Watts 1600 Watts
Efficiency 80 plus Platinum
Redundancy 1+1
Processor Q’ty / Socket Type (1) AMD Socket SP3
Supported CPU Series (1) AMD EPYC™ 7000 Series Processor
System Bus 1066/ 800/ 533 MHz
Average CPU Power (ACP) Wattage Max up to 180W
Memory Supported DIMM Qty (16) DIMM slots
DIMM Type / Speed DDR4 ECC RDIMM/LRDIMM 2666/2400
Capacity Up to 2,048GB RDIMM/ 1,024GB LRDIMM/ 2,048GB LRDIMM 3DS
*Actual memory speed depends on memory population topology
Memory channel 8 Channels per CPU
Memory voltage 1.2V
Expansion Slots PCI-E Default configuration: (4) PCI-E Gen3 x16 slots (Full-height, 10.5″ length, Double-width) + (1) PCI-E Gen3 x16 slot (Full-height, Half-length, Single-width)
Pre-install TYAN Riser Card (1) M2091-R riser card (right) for (1) PCI-E Gen3 x16 slot / (1) M8021-R16-1L-1 riser card for (1) PCI-E Gen3 x16 slot / (2) M5631-R16-1F-1 riser card for (2) PCI-E Gen3 x16 slots / (1) M8021-R32-2L-2 riser card for (1) PCI-E Gen3 x16 slot
Others: Field Upgradable Option:
PCI-E: (6) PCI-E Gen3 x16 slots (Full-height, 10.5″-length, Single-width) + (1) PCI-E Gen3 x16 slot (Full-height, Half-length, Single-width) / Pre-installed Riser Cards: (1) M2091-R riser card (right) for (1) PCI-E Gen3 x16 slot , (1) M8021-R32-2L-1 riser card for (2) PCI-E Gen3 x16 slots, (1) M8021-R32-2L-2 riser card for (2) PCI-E Gen3 x16 slots, (2) M5631-R16-1F-1 riser card for (2) PCI-E Gen3 x16 slots
LAN Q’ty / Port (2) 10GbE ports / (1) PHY
Controller Intel X550-AT2
PHY Realtek RTL8211E
Storage NVMe Connector (M.2) (2) 22110/2280 (by PCI-E interface)
Graphic Connector type D-Sub 15-pin
Resolution Up to 1920×1200
Chipset Aspeed AST2500
I/O Ports USB (2) USB3.0 ports (at front) / (1) USB3.0 ports (at rear)
COM (1) 2×12 pin-header
VGA (1) D-Sub 15-pin port
RJ-45 (2) 10GbE ports, (1) GbE dedicated for IPMI
TPM (Optional) TPM Support Please refer to our TPM supported list.
System Monitoring Chipset Aspeed AST2500
Temperature Monitors temperature for CPU & memory & system environment
Voltage Monitors voltage for CPU, memory, chipset & power supply
LED Over temperature warning indicator / Fan & PSU fail LED indicator
Others Watchdog timer support
Server Management Onboard Chipset Onboard Aspeed AST2500
AST2500 iKVM Feature 24-bit high quality video compression / Supports storage over IP and remote platform-flash / USB 2.0 virtual hub
AST2500 IPMI Feature IPMI 2.0 compliant baseboard management controller (BMC) / 10/100/1000 Mb/s MAC interface
BIOS Brand / ROM size AMI / 32MB
Feature Hardware Monitor / Boot from USB device/PXE via LAN/Storage / Console Redirection / ACPI 6.1 / SMBIOS 3.1/PnP/Wake on LAN / ACPI sleeping states S0, S5 / FAN speed control automatic
Operating System OS supported list Please refer to our AVL support lists.
Regulation FCC (SDoC) Class A
CE (DoC) Class A
CB/LVD Yes
RCM Class A
VCCI Class A
Operating Environment Operating Temp. 10° C ~ 35° C (50° F~ 95° F)
Non-operating Temp. – 40° C ~ 70° C (-40° F ~ 158° F)
In/Non-operating Humidity 90%, non-condensing at 35° C
RoHS RoHS 6/6 Compliant Yes
Package Contains Barebone (1) GA88-B8021 Barebone
Manual (1) Quick Installation Guide
Installation CD (1) TYAN Device Driver CD