Microsoft is integrating Nvidia\u2019s AI Enterprise software suite with its Azure Machine Learning service to help enterprise developers build, deploy, and manage applications based on large language models, it said Tuesday.\nDevelopers and enterprises will have access to over 100 frameworks, pretrained large language models, and development tools as part of AI Enterprise Suite integration with Microsoft\u2019s Azure Machine Learning service, the companies said in a joint statement. For now, the integration is only available through an invitation-only preview in the Nvidia community registry.\nNvidia\u2019s AI Enterprise Suite aids in accelerating the data science pipeline and streamlines development and deployment of production AI including generative AI, computer vision, and speech AI, the chip maker said.\nThe suite comes with tools such as Nvidia RAPIDS for accelerating data science workloads, Nvidia Metropolis for accelerating Vision AI model development, Nvidia Triton Inference Server for standardizing model deployment, and NeMo Guardrails software that enables developers to add safety and security features for AI chatbots, it added. Users will also have access to Nvidia experts and a support service.\nAs part of the collaboration between the two companies, Microsoft will make Nvidia\u2019s AI Enterprise software suite available on its Azure Marketplace.\nThe two companies are working to combine their software offerings in other areas too. Nvidia Omniverse Cloud platform-as-a-service (PaaS) is now available on Microsoft Azure as a private offer for enterprises. Omniverse Cloud provides developers and enterprises with a full-stack cloud environment to design, develop, deploy and manage industrial metaverse applications at scale, the companies said.\nIn the last few months, Nvidia has consistently partnered with several technology companies such as Oracle, Google Cloud, ServiceNow and Dell to provide services for developing AI and generative AI applications. And in March, the chip maker had said that it would make its DGX Pods, which power ChatGPT, available in the cloud.