Skip to content

🚨 Mission 01: Introduction to Agents

🕵️‍♂️ CODENAME: OPERATION AI AGENT DECODE

⏱️ Operation Time Window: ~30 minutes – intel only, no fieldwork required

🎥 Watch the Walkthrough

Introduction to Agents video thumbnail

🎯 Mission Brief

Welcome, Recruit. Before we dive into building agents, you need a firm grasp of the AI concepts that power them. This mission will equip you with foundational knowledge of conversational AI, large language models (LLMs), retrieval-augmented generation (RAG), and the types of agents you can create in Copilot Studio.

🔎 Objectives

In this mission, you’ll learn:

  1. What conversational AI is
  2. How Large Language Models (LLMs) power chat experiences
  3. What Retrieval-Augmented Generation (RAG) brings to the table
  4. The distinction between conversational agents and autonomous agents
  5. How agents in Copilot Studio leverage these concepts

Let’s dive in!


What Is Conversational AI?

Conversational AI refers to any system that can understand, process, and respond to human language (text or speech) in a way that feels natural. Think chatbots on websites that help you figure out where your order is or virtual personal assistants in your favorite apps. Under the hood, most modern conversational AIs rely on Large Language Models (LLMs).

Large Language Models (LLMs) 101

At the core of most conversational AI systems are Large Language Models, neural networks trained on massive amounts of text. These models learn the statistical patterns of language so they can generate coherent sentences, answer questions, brainstorm ideas or even create content. Key points to understand:

  1. Training Data: LLMs ingest terabytes of text (web pages, books, poems, articles). This “world knowledge” lets them respond on many topics.
  2. Tokenization: Text is broken into smaller units called tokens (words, subwords, or characters). The model predicts one token at a time.
  3. Context Window: Each LLM has a limit on how many tokens it can “see” at once. Beyond that limit, prior tokens get shortened.
  4. Prompting: You interact with an LLM by sending it a prompt (your question or request in a block of text). The better your prompt, the more focused and relevant the response from the LLM.

Pro Tip

A common analogy is that an LLM is like a “super-smart autocomplete.” It doesn’t truly understand meaning like a human brain, but it’s extremely good at predicting the next best word (or phrase) in a sequence.


Retrieval-Augmented Generation (RAG)

When LLMs rely solely on static training data, they might hallucinate or become outdated. RAG addresses this by letting the model “look up” fresh information before composing an answer. At a high level, RAG works like this:

  1. User Query: A user asks a question (e.g., “What’s the latest on Contoso’s quarterly earnings?”).
  2. Retriever Step: The system queries a knowledge source (documents, public websites, internal databases, SharePoint libraries, etc.) to find relevant information.
  3. Augmentation: Retrieved data gets appended to or prepended before sending to the LLM.
  4. Generation: The LLM ingests both the user’s question and the retrieved context, then generates a response that’s grounded in up-to-date data.

With RAG, your agent can call internal company wikis, APIs, or search an FAQ knowledge base—and return answers that aren’t limited to the static data that the model is trained on.


Conversational vs. Autonomous Agents

In the context of Copilot Studio, the term agent can refer to multiple flavors of AI assistants. It’s helpful to draw a line between:

Conversational Agents:

  • Requires two-way dialogue (text or speech) to work.
  • Persist context across multiple turns of a conversation.
  • Can hook into external tools or APIs (e.g., call a Power Automate flow, send calendar invites, manipulate data in Dataverse).
  • Ideal for customer support, FAQs, guided interactions or simple Q&A.
  • Examples:
    • An agent in Microsoft Teams that answers HR policy questions.
    • An agent on a public website that answer questions about your products.

Autonomous Agents:

  • Go beyond back-and-forth chat; they can kick off and take actions on behalf of the user.
  • Use LLM reasoning loops (think “plan → act → observe → replan”) to complete tasks.
  • Can also hook into external tools or APIs (e.g., call a Power Automate flow, send calendar invites, manipulate data in Dataverse).
  • Operate without constant human prompts. Once triggered, they can handle multi-step processes autonomously.
  • Examples:
    • An agent that generates a travel itinerary, books flights, and emails confirmations as soon as you put in a travel request in your backend system.
    • A “Meeting Summarizer” agent that joins a Teams call, transcribes it in real time, and writes an executive summary to OneNote.

Key Difference

Conversational agents wait for user input and require back and forth dialogue to work. Autonomous agents can execute based on external triggers and perform actions without any human interaction.


Agents in Copilot Studio

Copilot Studio unifies both conversational and autonomous scenarios under one framework. Here’s how Copilot Studio helps you build agents:

  1. Visual Agent Designer: A drag and drop canvas to build, test and deploy your agents.
  2. Model (LLM) Selection: Select from various AI models (OpenAI, Anthropic, Custom Models) to choose the best LLM for your agent scenario.
  3. Knowledge: Use out-of-the-box integrations for SharePoint, OneDrive, Dataverse, etc, enabling RAG out of the box.
  4. Tools: Hook into external tools or APIs to enable your agent to perform actions (e.g., call a Power Automate flow, send calendar invites, manipulate data in Dataverse)
  5. Multi-Modal Support: Copilot studio agents support file uploads and speech conversations.
  6. Publishing & Distribution: Once your agent is ready, you can publish it to Microsoft 365 Copilot, embed it on your website or choose from several other deployment channels.

🎉 Mission Complete

You’ve now completed your introduction to agents and foundational AI concepts. You understand:

  1. LLMs = The “Brain” of Your Agent
  2. Responsible for language understanding and generation.

  3. RAG = Real-Time Knowledge Integration

  4. Bridges the gap between a static LLM and ever-changing data sources.
  5. Retrieves and injects relevant documents or records into the LLM prompt.

  6. Conversational vs. Autonomous

  7. Conversational: Focus on back and forth dialogue
  8. Autonomous: No dialogue required, kicks off from external triggers and performs actions autonomously

Next up, you'll explore the fundamentals of Copilot Studio!

Stay sharp, Recruit - your AI journey is just beginning!

📚 Tactical Resources

đź”— Copilot Studio Documentation Home


Analytics