This example shows how to use Ollama
To run this example, you need to have an Ollama server running aside and have llama3:latest
model installed. For how to setup an Ollama server, please refer to Ollama.
Note
You can find the complete sample code here
Step 1: Install AutoGen.Ollama
First, install the AutoGen.Ollama package using the following command:
dotnet add package AutoGen.Ollama
For how to install from nightly build, please refer to Installation.
Step 2: Add using statement
using AutoGen.Core;
using AutoGen.Ollama.Extension;
Step 3: Create and chat OllamaAgent
In this step, we create an Ollama
using var httpClient = new HttpClient()
{
BaseAddress = new Uri("http://localhost:11434"),
};
var ollamaAgent = new OllamaAgent(
httpClient: httpClient,
name: "ollama",
modelName: "llama3:latest",
systemMessage: "You are a helpful AI assistant")
.RegisterMessageConnector()
.RegisterPrintMessage();
var reply = await ollamaAgent.SendAsync("Can you write a piece of C# code to calculate 100th of fibonacci?");