Skip to content

Latest commit

 

History

History
133 lines (100 loc) · 11.5 KB

index.md

File metadata and controls

133 lines (100 loc) · 11.5 KB
layout title nav_order permalink
home
SLM Innovator Lab
1
/

SLM Innovator Lab

{: .no_toc }

Agenda{: .btn .btn-primary .fs-5 .mb-4 .mb-md-0 .mr-2 } View it on GitHub{: .btn .fs-5 .mb-4 .mb-md-0 }

Table of contents

{: .no_toc .text-delta }

  1. TOC {:toc}

1. Overview

Unlock the full potential of your AI projects with the SLM Innovator Lab, powered by the Azure AIML Platform. Our lab is tailored for customers who excel in fine-tuning and deploying multiple SLM/LLM models on Azure, as well as those aiming to optimize base model performance through fine-tuning to create RAG applications. With the advanced capabilities of AI Studio, you can establish efficient and scalable LLMOps.

This hands-on lab is suitable for the following purposes:

  1. 1-day workshop (4-7 hours depending on customer) / 2-day workshop with LLMOps hands-on
  2. Hackathon starter code
  3. Reference guide for SLM fine-tuning&serving PoC/Prototype

2. Prerequisites

{: .note} Please do not forget to modify the .env file to match your account. Rename .env.sample to .env or copy and use it

Before starting, you have met the following requirements:

3. How to get started

  1. Create your compute instance in Azure ML. For code development, we recommend Standard_DS11_v2 (2 cores, 14GB RAM, 28GB storage, No GPUs).
  2. Open the terminal of the CI and run:
    git clone https://github.com/Azure/slm-innovator-lab.git
    cd slm-innovator-lab && conda activate azureml_py310_sdkv2
    pip install -r requirements.txt

4. Agenda

  • Background. Why Azure and Fine-tuning?: Before we get into the hands-on, we explain to participants exactly what the SLM Innovator Lab is all about and give them an overview of the GenAI ecosystem in Azure.

  • Lab 0. Requirements (Skip if you have already set up): If you are not familiar with the Azure environment, be sure to check it out!

  • Lab 1. Data preparation: Participants will dive into the critical first step of the GenAI pipeline. The session will focus on how to prepare data from real-world scenarios to create high-quality datasets necessary for fine-tuning models. Participants will learn how to process QnA (Questions & Answers) data and generate synthetic data to augment the training set, ensuring the model can handle a wide range of domain, including those in non-English languages. By the end of this lab, participants will have hands-on experience in transforming raw data into a format ready for effective AI model training.

  • Lab 2. Fine-tuning and serving: This lab guides participants through the process of fine-tuning SLMs and deploying them using Azure ML. The focus will be on simplifying the fine-tuning process, enabling participants to fine-tune pre-trained SLMs with their own datasets quickly and efficiently. The session will also demonstrate how to use Azure ML’s tools to serve these models as scalable APIs (Application Programming Interfaces), allowing them to be integrated into real-world applications with ease.

  • Lab 3. LLMOps: In this lab, participants will delve into the critical aspects of managing and optimizing SLMs within Azure AI Studio, with a particular emphasis on content safety and model evaluation. As organizations increasingly deploy GenAI models in production environments, ensuring that these models operate safely and effectively is paramount. This lab provides participants with the critical skills needed to ensure that their LLMs are not only technically robust but also safe and aligned with the ethical and operational standards of their organizations

  • Key takeaways: The final session of the program will be a reflective and strategic discussion focused on the next steps in the participants' GenAI journey. After gaining hands-on experience, participants will have the opportunity to share their insights, challenges, and aspirations. This session will emphasize the importance of listening to the customer’s voice, understanding their specific needs, and collaboratively defining the next actions to move towards successful Proof of Concept (PoC), Minimum Viable Product (MVP), and full production deployments. The goal is to ensure that each participant leaves with a clear roadmap for their AI projects, aligned with their business objectives and supported by the technical capabilities they have developed during the program.

5. Objectives

  • Platform Stickiness: When customers are tuning SLMs using Azure ML and AI Studio, they initially develop the SLM/LLMOps through AI Studio and tools like Prompt flow, LangChain, LlamaIndex, or Semantic Kernel. This forms the initial approach for the SLM Innovator Lab. However, the primary goal is to migrate compute workloads to an Azure-native platform. Specifically, this involves integrating Azure AKS (containerization with multi-GPU node pools), Azure Blob Storage, Azure SQL Database, Azure Data Lake, Microsoft Fabric, Azure Monitor and Log Analytics, while working closely with various technical support teams within Microsoft to address the anticipated challenges of launching into production. The hands-on experience shows the practical benefits of using Azure ML and Azure AI Studio, encouraging long-term commitment to the platform.

  • Ease of Use and Accessibility: The lab is designed to lower the barriers to entry for customers interested in SLM/LLM fine-tuning. It simplifies the setup process by providing a pre-configured environment that omits the need for complex configurations, making it easier for customers to get started quickly.

  • Earn trust: The lab is designed to build and solidify customer trust throughout the GenAI journey, from Proof of Concept (PoC) to Minimum Viable Product (MVP) and to Production. By deeply engaging with customers to understand their unique requirements and challenges, and by collaboratively navigating the complexities of data preparation, model fine-tuning, serving, and evaluation, the SLM Innovator Lab ensures that customers see tangible results and feel confident in the partnership. This trust is crucial for successfully launching GenAI projects, as it fosters stronger relationships and collaboration, ensuring that both the customer's goals and technical needs are met at every stage of the project. The lab's success in earning this trust is key to achieving long-term success and deeper integration in AIML projects.

References

Expand

Data preparation

SLM fine-tuning

Phi-3/Phi-3.5

Florence-2

LLMOps