Step-by-Step

Using Cohere Binary Embeddings in Azure AI Search and Command R/R+ Model via Azure AI Studio

In April 2024, we proudly announced our partnership with Cohere, allowing customers to seamlessly leverage Cohere models via the Azure AI Studio Model Catalog, as part of the Models as a Service (MaaS) offering. At Build 2024, Azure AI Search launched support for Binary Vectors. In this blog, we are excited to continue from our […]

Using Cohere Binary Embeddings in Azure AI Search and Command R/R+ Model via Azure AI Studio Continue Reading

Best Practices to Manage and Mitigate Security Recommendations

In the fast-evolving landscape of cloud security, Microsoft Defender for Cloud (MDC) stands as a robust Cloud Native Application Protection Platform (CNAPP). One of its standout features is the premium Cloud Security Posture Management (CSPM) solution, known as Defender CSPM. Among the myriads of advanced capabilities offered by Defender CSPM, the “Governance Rule” feature is

Best Practices to Manage and Mitigate Security Recommendations Continue Reading

Optimizing ETL Workflows: A Guide to Azure Integration and Authentication with Batch and Storage

Introduction When it comes to building a robust foundation for ETL (Extract, Transform, Load) pipelines, the trio of Azure Data Factory or Azure Synapse Analytics, Azure Batch, and Azure Storage is indispensable. These tools enable efficient data movement, transformation, and processing across diverse data sources, thereby helping us achieve our strategic goals. This document provides

Optimizing ETL Workflows: A Guide to Azure Integration and Authentication with Batch and Storage Continue Reading

Evaluating Large and Small Language Models on Custom Data Using Azure Prompt Flow

The Evolution of AI and the Challenge of Model Selection In recent years, the field of Artificial Intelligence (AI) has witnessed remarkable advancements, leading to the unprecedented surge in the development of small and large language models. They’re at the heart of various applications, aiding in everything from customer service chatbots to content creation and

Evaluating Large and Small Language Models on Custom Data Using Azure Prompt Flow Continue Reading

Azure AI Search now supports AI Vision multimodal and AI Studio embedding models

Keeping pace with AI representation learning requires continuous integration and adaptation to new advancements. In line with this, we’re excited to announce new updates to Azure AI Search‘s integrated vectorization (preview) feature. It now supports native multimodal search capabilities, that seamlessly manage both text and images during indexing and querying. Moreover, we’ve incorporated support for

Azure AI Search now supports AI Vision multimodal and AI Studio embedding models Continue Reading

Accelerate cloud security risk remediation with Microsoft Copilot for Security

As cloud environments experience rapid expansion, evolution, and increasing complexity, security teams face a significant and growing challenge in identifying, assessing, and remediating cloud security risks across multicloud environments and developer pipelines. With Copilot in Defender for Cloud, security teams can efficiently identify critical risks across their multicloud environments and developer pipelines and streamline remediation

Accelerate cloud security risk remediation with Microsoft Copilot for Security Continue Reading

Evaluate Small Language Models for RAG using Azure Prompt Flow (LLama3 vs Phi3)

Introduction: Recently, small language models have made significant progress in terms of quality and context size. These advancements have enabled new possibilities, making it increasingly viable to leverage these models for retrieval-augmented generation (RAG) use cases. Particularly in scenarios where cost sensitivity is a key consideration, small language models offer an attractive alternative.   This post

Evaluate Small Language Models for RAG using Azure Prompt Flow (LLama3 vs Phi3) Continue Reading

The LLM Latency Guidebook: Optimizing Response Times for GenAI Applications

Co-authors: Priya Kedia, Julian Lee, Manoranjan Rajguru, Shikha Agrawal, Michael Tremeer Contributors: Ranjani Mani, Sumit Pokhariyal, Sydnee Mayers Generative AI applications are transforming how we do business today, creating new, engaging ways for customers to engage with applications. However, these new LLM models require massive amounts of compute to run, and unoptimized applications can run

The LLM Latency Guidebook: Optimizing Response Times for GenAI Applications Continue Reading

Improving RAG performance with Azure AI Search and Azure AI prompt flow in Azure AI Studio

Content authored by: Arpita Parmar    Introduction If you’ve been delving into the potential of large language models (LLMs) for search and retrieval tasks, you’ve probably encountered Retrieval Augmented Generation (RAG) as a valuable technique. RAG enriches LLM-generated responses by integrating relevant contextual information, particularly when connected to private data sources. This integration empowers the

Improving RAG performance with Azure AI Search and Azure AI prompt flow in Azure AI Studio Continue Reading

Deploy a Gradio Web App on Azure with Azure App Service: a Step-by-Step Guide

A teaser image generated by DALL E 2 Context Gradio is an open-source Python package that you can use for free to create a demo or web app for your machine learning model, API, Azure AI Services integration or any Python function. You can run Gradio in Python notebooks or on a script. A Gradio

Deploy a Gradio Web App on Azure with Azure App Service: a Step-by-Step Guide Continue Reading