Nvidia used its GPU Technology Conference in San Jose to introduce new blade servers for on-premises use and announce new cloud AI acceleration.\nThe RTX Blade Server packs up to 40 Turing-generation GPUs into an 8U enclosure, and multiple enclosures can be combined into a "pod" with up to 1,280 GPUs working as a single system and using Mellanox technology as the storage and networking interconnect. Which likely explains why Nvidia is paying close to $7 billion for Mellanox.\nInstead of AI, where Nvidia has become a leader, the RTX Blade Server is positioned for 3D rendering, ray tracing and cloud gaming. The company said this setup will enable the rendering of realistic-looking 3D images in real time for VR and AR.\nDell EMC, HPE, Lenovo, ASUS and Supermicro were at GTC and all introduced RTX servers.\nOn the AI side of things, Nvidia introduced CUDA-X AI, which it claims is the world\u2019s only end-to-end acceleration library for data science. CUDA is Nvidia\u2019s language using a C++ syntax to specifically program its GPUs.\nThe typical workflow for deep learning, machine learning and data analytics is data processing, feature determination, training, verification and deployment. These are all very different steps in the process and typically require different types of processing. CUDA-X AI uses the NVIDIA Tensor Core GPUs to address the end-to-end AI pipeline.\nAnd it has considerable adoption out of the gate. CUDA-X AI has been adopted by all the major cloud services, like Amazon Web Services, Google Cloud Platform and Microsoft Azure, and it has been adopted by Charter, PayPal, SAS and Walmart.\nFor on-prem servers, Nvidia introduced a new generation of T4 GPU processors that CEO Jen-Hsun Huang said only draws 70 watts of power, a big reduction from the usual power hogging of GPUs, is \u201cthe size of a candy bar\u201d and it fits into every single one of the high-volume most popular data center servers in the world.\nAnd as is always the case, Nvidia announced major server vendor support. Cisco, Dell EMC, Fujitsu, HPE, Inspur, Lenovo and Sugon all now offer Nvidia T4 GPU servers for data analytics, machine learning and deep learning.\nIn addition, Amazon Web Services announced it will release its latest GPU-equipped instance with support for NVIDIA\u2019s T4 Tensor Core GPUs, with a focus on machine learning workloads. Amazon\u2019s Elastic Container Service for Kubernetes will support T4.\n\u201cBecause T4 GPUs are extremely efficient for AI inference, they are well-suited for companies that seek powerful, cost-efficient cloud solutions for deploying machine learning models into production,\u201d said Ian Buck, vice president and general manager and of accelerated computing at Nvidia,\u00a0in a blog post.