ai
985 TopicsAnnouncing the Microsoft AI Skills Fest: Save the date!
The best way to learn something new is by taking it one step at a time. We know all this talk of AI can be overwhelming, so how about we take it one skill at a time? At Microsoft, our mission has always been to create technology that empowers others to innovate and solve real-world problems. And it’s no different with AI—we want to help you learn to use this powerful technology to make your life easier, especially as it becomes an integral part of our daily lives. Sometimes, starting is the hardest part, so we want to make that part simple for you. This is why we’re excited to announce the Microsoft AI Skills Fest, a global event this April and May, designed to bring learners across the globe together to build their AI skills, from beginner explorers to the technologically gifted. Together, we can learn a new AI skill and maybe even set a groundbreaking record at the same time! Everyone everywhere is invited The AI Skills Fest is designed with you in mind, offering a wide variety of AI training for everyone, regardless of background or expertise. Join us to build your AI skills and unlock new opportunities for productivity, innovation, and growth. Tech professionals. Learn how to quickly build AI-powered solutions using Microsoft's AI apps and services. Gain skills and experience working with agents, AI security, Azure AI Foundry, GitHub Copilot, Microsoft Fabric, and more. Business professionals. Find out how much easier your work life can be when you use Microsoft Copilot to simplify tasks and let your creativity loose! Students. Explore technical skills to bring your ideas to life, with AI learning experiences for all skill levels and interests. Business leaders. Empower your teams with AI skills for future success through curated upskilling opportunities. Let’s earn a GUINNESS WORLD RECORDS™ title together on April 8, 2025! The Microsoft AI Skills Fest will begin with a spectacular Kickoff Celebration on April 8, 2025. Starting in Australia at 9 AM Australian Eastern Standard Time and wrapping up in the United States at 4 PM Pacific Daylight Time, this 24-hour, globe-spanning event will feature a variety of AI learning activities designed to engage and inspire learners of all experience levels. Together, we’ll have a once-in-a-lifetime opportunity to attempt a GUINNESS WORLD RECORDS™ title for most users to take an online multilevel artificial intelligence lesson in 24 hours. Don’t miss this unique chance to learn, compete, and celebrate your achievements—and to be part of these unprecedented and record-setting global festivities. Mark your calendar and save the date for this unparalleled event. Deepen your skills over 50 days of AI learning The celebration doesn't stop there! Following the kickoff, the AI Skills Fest will continue for 50 days— through May 28, 2025—offering a wide array of learning opportunities. Whether you're a tech professional, business leader, business professional, student, or AI enthusiast, there will be engaging and challenging training for you. Participate in hackathons, live in-person and virtual training, self-paced tutorials, on-demand training, community events, Microsoft Learn Challenges, and more. Mark your calendar for 50 days of AI discovery and learning. The countdown to make history together Save the date, attempt a world record, and get ready to unlock the future with 50 days of AI discovery and learning. Stay tuned for all the details on March 24, 2025. Now let’s get learning!3.5KViews17likes4CommentsGet certified as an Azure AI Engineer (AI-102) this summer?
For developers, the accreditation as an Azure AI Engineer—certified through the rigorous AI-102 exam—has become a golden ticket to career acceleration. It isn’t just about coding chatbots or fine-tuning machine learning models; it’s about gaining the confidence (for you and for your business) that you can wield Azure’s toolkits to configure AI solutions that augment human capability. Before we dive in, if you’re planning to become certified as an Azure AI Engineer, you may find this Starter Learning Plan (AI 102) valuable—recently curated by a group of Microsoft experts, purposed for your success. We recommend adding it to your existing learning portfolio. It’s a light introduction that should take less than four hours, but it offers a solid glimpse into what to expect on your journey and the breadth of solutions you might craft in the future. From revolutionizing customer service with intelligent agents to optimizing supply chains through predictive analytics, Azure AI engineers sit at the confluence of technological ingenuity and business transformation. For those with an appetite for problem-solving and a vision for AI-driven futures, this certification isn’t just another badge—it’s an assertion of expertise in a field where demand is outpacing supply. Securing that expertise, however, requires more than just a weekend of cramming. Today’s aspiring AI engineers navigate an ecosystem of learning that is as modern as the field itself. Gone are the days when one could rely solely on a stack of manuals; now, candidates immerse themselves in a medley of Microsoft Learn modules, hands-on labs, AI-powered coding assistants, and community-led study groups. Many take a pragmatic approach—building real-world projects using Azure Cognitive Services and Machine Learning Studio to cement their understanding. Others lean on practice exams and structured courses from platforms like Pluralsight and Udemy, ensuring they aren’t just memorizing but internalizing the core principles. The AI-102 exam doesn’t reward rote knowledge—it demands fluency in designing, deploying, and securing AI solutions, making thorough preparation an indispensable part of the journey. In addition to the above learning plan, we want to provide a few other tips. Understand the Exam Objectives: Begin by thoroughly reviewing the AI-102 study guide. This document outlines the key topics and skills assessed, including planning and managing Azure AI solutions, implementing computer vision and natural language processing solutions, and deploying generative AI solutions. Familiarizing yourself with these areas will provide a structured framework for your study plan. Continuous memorization is part of your study. But if you get a bit bored from your flashcards and look for more ‘storyline’ style learning content, we recommend adding MSFT employee created learning plan to your mix. They are scenario-based and focus more on providing you with a structured understanding of how to do XYZ on Azure. Here are 3 examples: Modernize for AI Readiness Build AI apps with Azure Re-platform AI applications Hands-On Practice: Practical experience is invaluable. Engage with Azure AI services directly by building projects that incorporate computer vision, natural language processing, and other AI functionalities. This hands-on approach not only reinforces theoretical knowledge but also enhances problem-solving skills in real-world scenarios. Utilize Practice Assessments: Assess your readiness by taking advantage of free practice assessments provided by Microsoft. These assessments mirror the style and difficulty of actual exam questions, offering detailed feedback and links to additional resources for areas that may require further study. Stay Updated on Exam Changes: Certification exams are periodically updated to reflect the latest technologies and practices. Regularly consult the official exam page to stay informed about any changes in exam content or structure. Participate in Community Discussions: Engaging with peers through forums and study groups can provide diverse perspectives and insights. The Microsoft Q&A platform is a valuable resource for asking questions, sharing knowledge, and learning from the experiences of others preparing for the same certification. By systematically incorporating these strategies into your preparation, you'll be well-positioned to excel in the AI-102 exam and advance your career as an Azure AI Engineer. If you have additional tips or thoughts, let us know in the comments area. Good luck!Use AI for Free with GitHub Models and TypeScript! 💸💸💸
Learn how to use AI for free with GitHub Models! Test models like GPT-4o without paying for APIs or setting up infrastructure. This step-by-step guide shows how to integrate GitHub Models with TypeScript in the Microblog AI Remix project. Start exploring AI for free today!Mondays at Microsoft | Episode 44
Busy last few weeks, with the SharePoint Hackathon and planning for the Microsoft 365 Community Conference. Stay in the know with Mondays at Microsoft. Karuana Gatimu and Heather Cook keep you grounded with the latest in #AI, broader Microsoft 365 product awareness, community activities and events, and more. Join live on Monday, March 10th, 8:00am PT. #CommunityLuv 💖 Resource links mentioned during the episode: (Note: Coming soon | We add all links and show notes at the completion of the live episode)368Views0likes26CommentsSpeed Up OpenAI Embedding By 4x With This Simple Trick!
In today’s fast-paced world of AI applications, optimizing performance should be one of your top priorities. This guide walks you through a simple yet powerful way to reduce OpenAI embedding response sizes by 75%—cutting them from 32 KB to just 8 KB per request. By switching from float32 to base64 encoding in your Retrieval-Augmented Generation (RAG) system, you can achieve a 4x efficiency boost, minimizing network overhead, saving costs and dramatically improving responsiveness. Let's consider the following scenario. Use Case: RAG Application Processing a 10-Page PDF A user interacts with a RAG-powered application that processes a 10-page PDF and uses OpenAI embedding models to make the document searchable from an LLM. The goal is to show how optimizing embedding response size impacts overall system performance. Step 1: Embedding Creation from the 10-Page PDF In a typical RAG system, the first step is to embed documents (in this case, a 10-page PDF) to store meaningful vectors that will later be retrieved for answering queries. The PDF is split into chunks. In our example, each chunk contains approximately 100 tokens (for the sake of simplicity), but the recommended chunk size varies based on the language and the embedding model. Assumptions for the PDF: - A 10-page PDF has approximately 3325 tokens (about 300 tokens per page). - You’ll split this document into 34 chunks (each containing 100 tokens). - Each chunk will then be sent to the embedding OpenAI API for processing. Step 2: The User Interacts with the RAG Application Once the embeddings for the PDF are created, the user interacts with the RAG application, querying it multiple times. Each query is processed by retrieving the most relevant pieces of the document using the previously created embeddings. For simplicity, let’s assume: - The user sends 10 queries, each containing 200 tokens. - Each query requires 2 embedding requests (since the query is split into 100-token chunks for embedding). - After embedding the query, the system performs retrieval and returns the most relevant documents (the RAG response). Embedding Response Size The OpenAI Embeddings models take an input of tokens (the text to embed) and return a list of numbers called a vector. This list of numbers represents the “embedding” of the input in the model so that it can be compared with another vector to measure similarity. In RAG, we use embedding models to quickly search for relevant data in a vector database. By default, embeddings are serialized as an array of floating-point values in a JSON document so each response from the embedding API is relatively large. The array values are 32-bit floating point numbers, or float32. Each float32 value occupies 4 bytes, and the embedding vector returned by models like OpenAI’s text-embedding-ada-002 typically consists of 1536-dimensional vectors. The challenge is the size of the embedding response: - Each response consists of 1536 float32 values (one per dimension). - 1536 float32 values result in 6144 bytes (1536 × 4 bytes). - When serialized as UTF-8 for transmission over the network, this results in approximately 32 KB per response due to additional serialization overhead (like delimiters). Optimizing Embedding Response Size One approach to optimize the embedding response size is to serialize the embedding as base64. This encoding reduces the overall size by compressing the data, while maintaining the integrity of the embedding information. This leads to a significant reduction in the size of the embedding response. With base64-encoded embeddings, the response size reduces from 32 KB to approximately 8 KB, as demonstrated below: base64 vs float32 Min (Bytes) Max (Bytes) Mean (Bytes) Min (+) Max (+) Mean (+) 100 tokens embeddings: text-embedding-3-small 32673.000 32751.000 32703.800 8192.000 (4.0x) (74.9%) 8192.000 (4.0x) (75.0%) 8192.000 (4.0x) (74.9%) 100 tokens embeddings: text-embedding-3-large 65757.000 65893.000 65810.200 16384.000 (4.0x) (75.1%) 16384.000 (4.0x) (75.1%) 16384.000 (4.0x) (75.1%) 100 tokens embeddings: text-embedding-ada-002 32882.000 32939.000 32909.000 8192.000 (4.0x) (75.1%) 8192.000 (4.0x) (75.2%) 8192.000 (4.0x) (75.1%) The source code of this benchmark can be found at: https://github.com/manekinekko/rich-bench-node (kudos to Anthony Shaw for creating the rich-bench python runner) Comparing the Two Scenarios Let’s break down and compare the total performance of the system in two scenarios: Scenario 1: Embeddings Serialized as float32 (32 KB per Response) Scenario 2: Embeddings Serialized as base64 (8 KB per Response) Scenario 1: Embeddings Serialized as Float32 In this scenario, the PDF embedding creation and user queries involve larger responses due to float32 serialization. Let’s compute the total response size for each phase: 1. Embedding Creation for the PDF: - 34 embedding requests (one per 100-token chunk). - 34 responses with 32 KB each. Total size for PDF embedding responses: 34 × 32 KB = 1088 KB = 1.088 MB 2. User Interactions with the RAG App: - Each user query consists of 200 tokens (which is split into 2 chunks of 100 tokens). - 10 user queries, requiring 2 embedding responses per query (for 2 chunks). - Each embedding response is 32 KB. Total size for user queries: Embedding responses: 20 × 32 KB = 640 KB. RAG responses: 10 × 32 KB = 320 KB. Total size for user interactions: 640 KB (embedding) + 320 KB (RAG) = 960 KB. 3. Total Size: Total size for embedding responses (PDF + user queries): 1088 KB + 640 KB = 1.728 MB Total size for RAG responses: 320 KB. Overall total size for all 10 responses: 1728 KB + 320 KB = 2048 KB = 2 MB Scenario 2: Embeddings Serialized as Base64 In this optimized scenario, the embedding response size is reduced to 8 KB by using base64 encoding. 1. Embedding Creation for the PDF: - 34 embedding requests. - 34 responses with 8 KB each. Total size for PDF embedding responses: 34 × 8 KB = 272 KB. 2. User Interactions with the RAG App: - Embedding responses for 10 queries, 2 responses per query. - Each embedding response is 8 KB. Total size for user queries: Embedding responses: 20 × 8 KB = 160 KB. RAG responses: 10 × 8 KB = 80 KB. Total size for user interactions: 160 KB (embedding) + 80 KB (RAG) = 240 KB 3. Total Size (Optimized Scenario): Total size for embedding responses (PDF + user queries): 272 KB + 160 KB = 432 KB. Total size for RAG responses: 80 KB. Overall total size for all 10 responses: 432 KB + 80 KB = 512 KB Performance Gain: Comparison Between Scenarios The optimized scenario (base64 encoding) is 4 times smaller than the original (float32 encoding): 2048 / 512 = 4 times smaller. The total size reduction between the two scenarios is: 2048 KB - 512 KB = 1536 KB = 1.536 MB. And the reduction in data size is: (1536 / 2048) × 100 = 75% reduction. How to Configure base64 encoding format When getting a vector representation of a given input that can be easily consumed by machine learning models and algorithms, as a developer, you usually call either the OpenAI API endpoint directly or use one of the official libraries for your programming language. Calling the OpenAI or Azure OpenAI APIs Using OpenAI endpoint: curl -X POST "https://api.openai.com/v1/embeddings" \ -H "Content-Type: application/json" \ -H "Authorization: Bearer YOUR_API_KEY" \ -d '{ "input": "The five boxing wizards jump quickly", "model": "text-embedding-ada-002", "encoding_format": "base64" }' Or, calling Azure OpenAI resources: curl -X POST "https://{endpoint}/openai/deployments/{deployment-id}/embeddings?api-version=2024-10-21" \ -H "Content-Type: application/json" \ -H "api-key: YOUR_API_KEY" \ -d '{ "input": ["The five boxing wizards jump quickly"], "encoding_format": "base64" }' Using OpenAI Libraries JavaScript/TypeScript const response = await client.embeddings.create({ input: "The five boxing wizards jump quickly", model: "text-embedding-3-small", encoding_format: "base64" }); A pull request has been sent to the openai SDK for Node.js repository to make base64 the default encoding when/if the user does not provide an encoding. Please feel free to give that PR a thumb up. Python embedding = client.embeddings.create( input="The five boxing wizards jump quickly", model="text-embedding-3-small", encoding_format="base64" ) NB: from 1.62 the openai SDK for Python will default to base64. Java EmbeddingCreateParams embeddingCreateParams = EmbeddingCreateParams .builder() .input("The five boxing wizards jump quickly") .encodingFormat(EncodingFormat.BASE64) .model("text-embedding-3-small") .build(); .NET The openai-dotnet library is already enforcing the base64 encoding, and does not allow setting encoding_format by the user (see). Conclusion By optimizing the embedding response serialization from float32 to base64, you achieved a 75% reduction in data size and improved performance by 4x. This reduction significantly enhances the efficiency of your RAG application, especially when processing large documents like PDFs and handling multiple user queries. For 1 million users sending 1,000 requests per month, the total size saved would be approximately 22.9 TB per month simply by using base64 encoded embeddings. As demonstrated, optimizing the size of the API responses is not only crucial for reducing network overhead but also for improving the overall responsiveness of your application. In a world where efficiency and scalability are key to delivering robust AI-powered solutions, this optimization can make a substantial difference in both performance and user experience. ■ Shoutout to my colleague Anthony Shaw for the the long and great discussions we had about embedding optimisations.Unlocking the Power of Azure Container Apps in 1 Minute Video
Azure Container Apps provides a seamless way to build, deploy, and scale cloud-native applications without the complexity of managing infrastructure. Whether you’re developing microservices, APIs, or AI-powered applications, this fully managed service enables you to focus on writing code while Azure handles scalability, networking, and deployments. In this blog post, we explore five essential aspects of Azure Container Apps—each highlighted in a one-minute video. From intelligent applications and secure networking to effortless deployments and rollbacks, these insights will help you maximize the capabilities of serverless containers on Azure. Azure Container Apps - in 1 Minute Azure Container Apps is a fully managed platform designed for cloud-native applications, providing effortless deployment and scaling. It eliminates infrastructure complexity, letting developers focus on writing code while Azure automatically handles scaling based on demand. Whether running APIs, event-driven applications, or microservices, Azure Container Apps ensures high performance and flexibility with minimal operational overhead. Watch the video on YouTube Intelligent Apps with Azure Container Apps – in 1 Minute Azure Container Apps, Azure OpenAI, and Azure AI Search make it possible to build intelligent applications with Retrieval-Augmented Generation (RAG). Your app can call Azure OpenAI in real-time to generate and interpret data, while Azure AI Search retrieves relevant information, enhancing responses with up-to-date context. For advanced scenarios, AI models can execute live code via Azure Container Apps, and GPU-powered instances support fine-tuning and inferencing at scale. This seamless integration enables AI-driven applications to deliver dynamic, context-aware functionality with ease. Watch the video on YouTube Networking for Azure Container Apps: VNETs, Security Simplified – in 1 Minute Azure Container Apps provides built-in networking features, including support for Virtual Networks (VNETs) to control service-to-service communication. Secure internal traffic while exposing public endpoints with custom domain names and free certificates. Fine-tuned ingress and egress controls ensure that only the right traffic gets through, maintaining a balance between security and accessibility. Service discovery is automatic, making inter-app communication seamless within your Azure Container Apps environment. Watch the video on YouTube Azure Continuous Deployment and Observability with Azure Container Apps - in 1 Minute Azure Container Apps simplifies continuous deployment with built-in integrations for GitHub Actions and Azure DevOps pipelines. Every code change triggers a revision, ensuring smooth rollouts with zero downtime. Observability is fully integrated via Azure Monitor, Log Streaming, and the Container Console, allowing you to track performance, debug live issues, and maintain real-time visibility into your app’s health—all without interrupting operations. Watch the video on YouTube Effortless Rollbacks and Deployments with Azure Container Apps – in 1 Minute With Azure Container Apps, every deployment creates a new revision, allowing multiple versions to run simultaneously. This enables safe, real-time testing of updates without disrupting production. Rolling back is instant—just select a previous revision and restore your app effortlessly. This powerful revision control system ensures that deployments remain flexible, reliable, and low-risk. Watch the video on YouTube Watch the Full Playlist For a complete overview of Azure Container Apps capabilities, watch the full JavaScript on Azure Container Apps YouTube Playlist Create Your Own AI-Powered Video Content Inspired by these short-form technical videos? You can create your own AI-generated videos using Azure AI to automate scriptwriting and voiceovers. Whether you’re a content creator, or business looking to showcase technical concepts, Azure AI makes it easy to generate professional-looking explainer content. Learn how to create engaging short videos with Azure AI by following our open-source AI Video Playbook. Conclusion Azure Container Apps is designed to simplify modern application development by providing a fully managed, serverless container environment. Whether you need to scale microservices, integrate AI capabilities, enhance security with VNETs, or streamline CI/CD workflows, Azure Container Apps offers a comprehensive solution. By leveraging its built-in features such as automatic scaling, revision-based rollbacks, and deep observability, developers can deploy and manage applications with confidence. These one-minute videos provide a quick technical overview of how Azure Container Apps empowers you to build scalable, resilient applications with ease. FREE Content Check out our other FREE content to learn more about Azure services and Generative AI: Generative AI for Beginners - A JavaScript Adventure! Learn more about Azure AI Agent Service LlamaIndex on Azure JavaScript on Azure Container Apps JavaScript at MicrosoftConstruyendo una Aplicación Web con Inteligencia Artificial usando Python
En la segunda sesión del GitHub Copilot Bootcamp LATAM, organizado por Microsoft Reactor, el ingeniero Manuel Ortiz, Embajador de Microsoft Learn y líder comunitario en GitHub, guió a desarrolladores en la creación de una aplicación web con capacidades de inteligencia artificial. Este taller práctico combinó fundamentos de desarrollo backend en Python con técnicas avanzadas de integración de modelos de lenguaje de Azure OpenAI. Introducción a Azure Open AI Azure Open AI es una colaboración entre Microsoft y OpenAI que permite a los desarrolladores integrar modelos avanzados de inteligencia artificial en sus aplicaciones utilizando la infraestructura de Azure. Esto ofrece acceso a modelos poderosos como GPT-4, que pueden ser utilizados para una variedad de tareas, desde procesamiento de lenguaje natural hasta generación de texto. Configuración de Azure Open AI Para comenzar a usar Azure Open AI, debes seguir algunos pasos básicos: Crear una Cuenta en Azure: Si aún no tienes una cuenta, puedes crear una en el portal de Azure. Los estudiantes pueden solicitar créditos gratuitos para usar los servicios de Azure. Crear un Servicio Azure Open AI: Accede al portal de Azure y busca "Azure Open AI". Haz clic en "Crear" y selecciona tu suscripción y grupo de recursos. Elige la región y configura el nombre del servicio, que debe ser alfanumérico y sin caracteres especiales. Selecciona el plan de precios adecuado y finaliza la creación del servicio. Obtener las Credenciales: Después de crear el servicio, necesitarás las credenciales (clave de API y endpoint) para autenticar tus solicitudes. Estas credenciales se pueden encontrar en la sección de "Claves y Endpoints" del servicio creado. Integración con Python y Flask Python es uno de los lenguajes de programación más populares para el desarrollo de aplicaciones de inteligencia artificial debido a su simplicidad y vasta biblioteca de herramientas. Durante la configuración, puedes usar varias bibliotecas y herramientas que facilitan el desarrollo de IA con Python, incluyendo: TensorFlow: Una biblioteca de código abierto para aprendizaje automático. Keras: Una API de alto nivel para redes neuronales, que funciona sobre TensorFlow. Scikit-learn: Una biblioteca para aprendizaje automático en Python. Flask: Un microframework para desarrollo de aplicaciones web. Una vez configurado el servicio Azure Open AI, puedes integrarlo en tus aplicaciones Python usando Flask. Aquí tienes un ejemplo de cómo hacerlo: Instalación de las Bibliotecas Necesarias: Crea un entorno virtual e instala las bibliotecas necesarias, como flask y openai. Configuración del Proyecto: Crea un archivo .env para almacenar tus credenciales de forma segura. Configura tu aplicación Flask para cargar estas credenciales y conectarse al servicio Azure Open AI. Creación del Modelo de IA: Utiliza la biblioteca openai para enviar prompts al modelo y recibir respuestas. Integra estas respuestas en tu aplicación web para proporcionar funcionalidades de IA a los usuarios. Ejemplo de Código Aquí tienes un ejemplo simplificado de cómo configurar y usar Azure Open AI en una aplicación Flask: from flask import Flask, request, render_template import openai import os app = Flask(__name__) # Cargar las credenciales del archivo .env openai.api_key = os.getenv("AZURE_OPENAI_API_KEY") openai.api_base = os.getenv("AZURE_OPENAI_ENDPOINT") @app.route("/", methods=["GET", "POST"]) def index(): response_text = "" if request.method == "POST": prompt = request.form["prompt"] response = openai.Completion.create( engine="text-davinci-003", prompt=prompt, max_tokens=100 ) response_text = response.choices.text.strip() return render_template("index.html", response_text=response_text) if __name__ == "__main__": app.run(debug=True) Beneficios de Azure Open AI Acceso a Modelos Avanzados: Utiliza los modelos más recientes y poderosos de OpenAI. Escalabilidad: La infraestructura de Azure permite escalar tus aplicaciones según sea necesario. Seguridad y Conformidad: Benefíciate de las robustas medidas de seguridad y conformidad de Azure. Sigue aprendiendo Si deseas aprender más sobre estas técnicas, mira las grabaciones del GitHub Copilot Bootcamp, comienza a utilizar el GitHub Copilot gratuito y descubre cómo transformar tu manera de programar utilizando inteligencia artificial.