Blog

Microsoft Azure AI Service Introduces GPT-4 and New Virtual Machine Series For Generative AI

Microsoft Azure AI Service- Featured img BDCC

The wait is over! Microsoft continues its commitment towards Responsible AI by releasing GPT Version 4 in its OpenAI Service Platform. Microsoft has been going hand-in-hand with OpenAI since the global release of ChatGPT. Microsoft Azure AI Services now has embedded GPT 4 alongside its existing Generative AI models.

“Azure OpenAI Service comes with built-in Generative AI models and collaborates with OpenAI’s ChatGPT, GPT-3.5/4, and DALL•E 2 to offer enterprise-readiness for various use cases”.

Microsoft is always at the forefront of adapting Generative AI transformation. With the availability of GPT 4 on Azure OpenAI, Microsoft has introduced an exclusive line-up of powerful Virtual Machines for supporting the changing demands of Generative AI models. This blog post details the two latest updates to the Azure OpenAI service – The release of GPT 4 and the latest VM services to support the hardware needs of Generative AI.

OpenAI GPT 4 Is Now Available In Microsoft Azure AI Service

Microsoft Azure users can now access and use OpenAI GPT 4 using Azure OpenAI Studio. Based on Microsoft’s update about the billing for GPT 4 usage, the pricing is as below:

  • For eight thousand chat contexts, each prompt and each completion will cost $0.03 and $0.06 for 1000 tokens, respectively.
  • For thirty-two thousand chat contexts, each prompt and each completion will cost $0.06 and $0.12 for 1000 tokens, respectively.

GPT Version 4 can potentially take the Generative AI experience to another level using its broader domain intelligence and problem-solving abilities. Businesses of all types can benefit from using GPT-4 in Azure. From streamlining internal communications to using a large-scale AI model, companies can utilize Microsoft AI Services to leverage different language models using Azure OpenAI Service.

Quickstart: Using GPT 4 with Azure OpenAI Studio

Prerequisites:

  • Access to Azure OpenAI Studio with an active subscription
  • An OpenAI Service Resource with a pre-deployed GPT-35-Turbo/GPT4 model

First, visit Azure OpenAI Studio with the link https://oai.azure.com/ and sign in to Azure. Now, here is the instruction to use GPT 4 in Azure:

  • Assistant Setup: Use pre-loaded System message examples and add few-shot examples to enhance in-context learning. You can choose ‘Xbox customer support agent’ from the dropdown and continue.
  • Chat Session: Send messages to the completions API for responses in the Microsoft Azure AI chat option. Engage in a conversation and get prompts based on your input. You can clear the chat anytime to start anew.
  • Settings: Tailor your deployment with temperature, max length, and more parameters. Learn about system messages to understand assistant responses.
  • Deploying GPT 4 Models: Use Microsoft OpenAI to deploy your model as a web app using the ‘Deploy to’ option. You can choose a new or existing web app and set app details for future usage.

Instructions: Join Azure OpenAI Waitlist To Use GPT 4

GPT-4 models are currently not available for public usage. If you want to access these models, you must join the waitlist. Azure has the form online for interested requestors. The form asks a few random questions to understand the purpose of usage and your intestine behind trying out GPT 4 in OpenAI Service. You can also submit different use cases using the same form. Once your GPT 4 access gets granted, you will get confirmation.

Introducing Microsoft’s Latest Virtual Machine Series To Power Generative AI Models

Azure Virtual Machine with extreme capabilities backs the Generative AI models of Azure OpenAI Service. So Far, Microsoft has introduced various Virtual Machines series. However, Generative AI models require extremely powerful and highly predictive computing power. That is why Microsoft Azure AI has launched the ND H100 v5 VM series. The top capabilities of this VM series include the following:

  • Interconnected 8xNVIDIA H100 Tensor GPUs with NextGen NVLink 4.0 and NVSwitch
  • NextGen NVLink 4.0 and NVSwitch with 3.6TB/s bisectional bandwidth for each VM
  • Interconnection with the NVIDIA Quantum-2 InfiniBand networking
  • 400GB per second NVIDIA Quantum-2 CX7 InfiniBand/GPU with 3.2TB per second per VM
  • 4th Gen Intel Xeon processors with 16 Channels of 4800MHz DDR5 DIMMs
  • Facility to connect PCIE Gen5 host with GPU of 64GB/s bandwidth

Microsoft’s new NDv5 H100 Virtual Machine Series is available for enterprise usage. Microsoft plans to assist enterprises that need high computing power to support their Generative AI applications.

How will NDv5 H100 Virtual Machine Series help in large-scale AI Models Deployment?

As generative AI gains momentum across diverse industries, the New Virtual Machines from the Series NDv5 H100 usher in a new era for large-scale AI model deployment. Microsoft’s expertise in delivering ExaOP Machines to Azure customers ensures exceptional performance. The new VMs result from efficient collaboration between Inflection, NVIDIA, and OpenAI.

The NDv5 H100 Virtual Machine Series empowers the creation of complex language models. Mustafa Suleyman, CEO of Inflection, appraises Azure’s AI infrastructure for enhancing the development of intricate AI models. With Azure’s pioneering AI supercomputer and services like Azure Machine Learning, large-scale generative AI models become accessible to startups and enterprises. NVIDIA’s Ian Buck highlights NDv5 H100’s pivotal role in driving generative AI applications and Azure DevOps services.

Highlighting The Features Of ND H100 v5 VM Series

  • Next-generation CPU: CPU performance matters the most for AI training models. This VM series comes with the 4th Gen Intel Xeon CPU as the foundation for producing optimal processing speed.
  • Supercomputing GPUs: This VM series has eight NVIDIA-H100 Tensor Core GPUs that boost the performance of AI models with unmatched computational power.
  • Optimized Performance: These VMS have Gen5 64GB/s bandwidth for each GPU, enhancing the host-to-GPU performance ratio.
  • Low-latency networking: This series has the networking power of NVIDIA Quantum-2 InfiniBand with 3.2 Tb/s per VM bandwidth across the GPUs.
  • Large-scale memory: The VMs in this Series has DDR5 memory that delivers smooth data transfer speeds and handles heavy workloads with larger datasets effortlessly.

NVIDIA Teams With Microsoft: Building Massive Cloud AI Computer In Progress

NVIDIA’s collaboration with Microsoft Azure AI is shaping a colossal Cloud AI computer that harnesses Azure’s potent infrastructure and NVIDIA GPUs. This venture establishes a remarkable platform by incorporating state-of-the-art AI capabilities on Microsoft Azure. The convergence of NVIDIA’s advanced AI stack and Azure’s scalable virtual machine instances propels large-scale AI models’ development, training, and deployment.

Rewinding Microsoft’s Promise: A Responsible Approach For Generative AI Development

With the global release of Azure OpenAI Service, Microsoft brings cutting-end Generative AI models, including GPT-35-Turbo and GPT 4, for enterprises worldwide. Additionally, Microsoft has introduced a series of highly powerful Virtual Machines to back these Generative AI models’ computing and hardware requirements.

However, Microsft’s promise to cater to Responsible AI remains at the core of Microsoft Azure AI Services. The AI Platform has built-in safety systems to tackle harmful inputs/prompts responsibly. Further, Microsoft assures the maximum protection of user data and privacy. Businesses can easily access the Azure OpenAI Studio to explore its features and capabilities. So it’s time to unlock the true power of Generative AI using Azure OpenAI Service!

FAQs

#1 How can I access Azure OpenAI Service?

With the general availability of Azure OpenAI Service, you can use the service with an active Azure Subscription. First, you need an Azure Account. Then you can register for OpenAI online. Microsoft may limit your access based on your location and service availability. Click on the form to request access.

#2 What is the pricing model for Azure OpenAI Service?

Pricing of most Azure services follows the consumption model of “pay-as-you-go.” However, the pricing model for Azure OpenAI Service is a bit different. The pricing depends on per unit or prompts for each Generative AI model in Microsoft Azure AI. Check the Azure OpenAI Service pricing options here.

#3 What are the available Generative AI Models through Azure OpenAI Service?

Azure OpenAI has a diverse range of Generative AI models with different capabilities, including the following:

  • GPT 4 models bring improvements from the old 3.5 version and generate natural language responses.
  • GPT 3.5 models that can process natural languages and programming code.
  • Embeddings that can convert human inputs into numerical vectors.
  • DALL-E models that can generate pictures based on human inputs.

#4 How do the Azure OpenAI capabilities compare to OpenAI?

Azure OpenAI Service from Microsoft incorporates the Generative AI models of OpenAI to offer customers advanced AI-based language processing capabilities. You can take the suggestions of Top Azure Consultants like Algoworks to understand the key difference between the two. However, Azure OpenAI not only incorporates OpenAI models like GPT but also incorporates Codex and DALL-E models.

#5 Does Azure OpenAI support Microsoft’s promise for Responsible AI?

Azure OpenAI follows Microsoft’s commitment to safeguard your data while using Azure OpenAI Service. It further implements the security measures on the available models in Azure OpenAI. It filters out biased/tampered content to offer users accurate information responsibly.

The following two tabs change content below.
BDCC

BDCC

Co-Founder & Director, Business Management
BDCC Global is a leading DevOps research company. We believe in sharing knowledge and increasing awareness, and to contribute to this cause, we try to include all the latest changes, news, and fresh content from the DevOps world into our blogs.
BDCC

About BDCC

BDCC Global is a leading DevOps research company. We believe in sharing knowledge and increasing awareness, and to contribute to this cause, we try to include all the latest changes, news, and fresh content from the DevOps world into our blogs.