Categories: EnterpriseSoftware

NVIDIA Launches AI Data Center Platform

At GTC in Tokyo, NVIDIA announced the launch of its new AI data center platform, the NVIDIA TensorRT Hyperscale Platform. This new platform is stated as delivering the most advanced inference accelerator for voice, video, image, and recommendation services. The platform will leverage the company’s new Tesla T4 GPUs as well as a comprehensive set of new inference software.


At GTC in Tokyo, NVIDIA announced the launch of its new AI data center platform, the NVIDIA TensorRT Hyperscale Platform. This new platform is stated as delivering the most advanced inference accelerator for voice, video, image, and recommendation services. The platform will leverage the company’s new Tesla T4 GPUs as well as a comprehensive set of new inference software.

Data centers process all types of queries now including voice, translations, images, videos, and various social media interactions. In order to address all of these different quires—that will require different types of neural networks—organizations need to leverage AI. NVIDIA’s new TensorRT Hyperscale Platform is a combination of hardware and software aimed at addressing the above issues. Leveraging Tesla T4 GPUs, based on the company’s Turing architecture, the new platform will be able to deliver high-performance with low latency for end-to-end applications. 

Key elements include:

  • NVIDIA Tesla T4 GPU – Featuring 320 Turing Tensor Cores and 2,560 CUDA cores, this new GPU provides breakthrough performance with flexible, multi-precision capabilities, from FP32 to FP16 to INT8, as well as INT4. Packaged in an energy-efficient, 75-watt, small PCIe form factor that easily fits into most servers, it offers 65 teraflops of peak performance for FP16, 130 teraflops for INT8 and 260 teraflops for INT4.
  • NVIDIA TensorRT 5 – An inference optimizer and runtime engine, NVIDIA TensorRT 5 supports Turing Tensor Cores and expands the set of neural network optimizations for multi-precision workloads.
  • NVIDIA TensorRT inference server – This containerized microservice software enables applications to use AI models in data center production. Freely available from the NVIDIA GPU Cloud container registry, it maximizes data center throughput and GPU utilization, supports all popular AI models and frameworks, and integrates with Kubernetes and Docker.

NVIDIA TensorRT Hyperscale Platform

Discuss this story

Sign up for the StorageReview newsletter

Adam Armstrong

Adam is the chief news editor for StorageReview.com, managing our internal and freelance content teams.

Recent Posts

Dell Advances Data Protection Portfolio Amid Rising Cyber Threats

Dell Technologies is advancing its data protection portfolio to enhance cyber resiliency across appliances, software, and as-a-service offerings amid rising…

2 days ago

HPE Cray Storage Systems C500 Lowers Storage Costs For Entry-level Snd Midrange HPC/AI Clusters

Since its launch in 2019, the Cray ClusterStor E1000 Storage System has emerged as a pivotal technology in the field…

2 days ago

Quantum Introduces Quantum GO Subscription Service For Data Management

Quantum Corporation has introduced Quantum GO, a subscription service designed to meet the escalating data demands and cost considerations enterprises…

3 days ago

JetCool Unveils Cold Plates for the NVIDIA H100 GPU

JetCool has launched an innovative liquid cooling module tailored for NVIDIA's H100 SXM and PCIe GPUs, claiming a significant advancement…

5 days ago

iXsystems Expands TrueNAS Enterprise with H-Series Platforms

iXsystems has launched the TrueNAS Enterprise H-Series platforms, designed to give organizations ultimate performance. The H10 model is now available,…

1 week ago

Microsoft Azure Edge Infrastructure At Hannover Messe 2024

Hannover Messe 2024 represents a significant event in the global industrial sector, serving as the world's largest industrial trade fair.…

1 week ago