The integration of Nvidia’s software suite will further help enterprises build, deploy, and manage applications based on large language models. Microsoft is integrating Nvidia’s AI Enterprise software suite with its Azure Machine Learning service to help enterprise developers build, deploy, and manage applications based on large language models, it said Tuesday. Developers and enterprises will have access to over 100 frameworks, pretrained large language models, and development tools as part of AI Enterprise Suite integration with Microsoft’s Azure Machine Learning service, the companies said in a joint statement. For now, the integration is only available through an invitation-only preview in the Nvidia community registry. Nvidia’s AI Enterprise Suite aids in accelerating the data science pipeline and streamlines development and deployment of production AI including generative AI, computer vision, and speech AI, the chip maker said. The suite comes with tools such as Nvidia RAPIDS for accelerating data science workloads, Nvidia Metropolis for accelerating Vision AI model development, Nvidia Triton Inference Server for standardizing model deployment, and NeMo Guardrails software that enables developers to add safety and security features for AI chatbots, it added. Users will also have access to Nvidia experts and a support service. As part of the collaboration between the two companies, Microsoft will make Nvidia’s AI Enterprise software suite available on its Azure Marketplace. The two companies are working to combine their software offerings in other areas too. Nvidia Omniverse Cloud platform-as-a-service (PaaS) is now available on Microsoft Azure as a private offer for enterprises. Omniverse Cloud provides developers and enterprises with a full-stack cloud environment to design, develop, deploy and manage industrial metaverse applications at scale, the companies said. In the last few months, Nvidia has consistently partnered with several technology companies such as Oracle, Google Cloud, ServiceNow and Dell to provide services for developing AI and generative AI applications. And in March, the chip maker had said that it would make its DGX Pods, which power ChatGPT, available in the cloud. Related content brandpost Sponsored by HPE Aruba Networking Bringing the data processing unit (DPU) revolution to your data center By Mark Berly, CTO Data Center Networking, HPE Aruba Networking Dec 04, 2023 4 mins Data Center feature 5 ways to boost server efficiency Right-sizing workloads, upgrading to newer servers, and managing power consumption can help enterprises reach their data center sustainability goals. By Maria Korolov Dec 04, 2023 9 mins Green IT Servers Data Center news Omdia: AI boosts server spending but unit sales still plunge A rush to build AI capacity using expensive coprocessors is jacking up the prices of servers, says research firm Omdia. By Andy Patrizio Dec 04, 2023 4 mins CPUs and Processors Generative AI Data Center feature What is Ethernet? History, evolution and roadmap The Ethernet protocol connects LANs, WANs, Internet, cloud, IoT devices, Wi-Fi systems into one seamless global communications network. By John Breeden Dec 04, 2023 11 mins Networking Podcasts Videos Resources Events NEWSLETTERS Newsletter Promo Module Test Description for newsletter promo module. Please enter a valid email address Subscribe