Module ai.ollama
ballerinax/ai.ollama Ballerina library
1.2.2
Overview
Ollama allows you to run open-source large language models (LLMs), such as Llama 3, Mistral, and Gemma, locally.
The Ollama connector offers APIs for connecting with locally running Ollama models, enabling the integration of advanced conversational AI and language processing capabilities into applications.
Key Features
- Connect and interact with local Ollama Large Language Models (LLMs)
- Support for Llama 3, Mistral, Gemma, and other open-source models
- Efficient handling of local model prompts and completions
- Streamlined integration with local AI infrastructure
Prerequisites
Ensure that your Ollama server is running locally before using this module in your Ballerina application.
Quickstart
To use the ai.ollama module in your Ballerina application, update the .bal file as follows:
Step 1: Import the module
Import the ai.ollama module.
import ballerinax/ai.ollama;
Step 2: Intialize the Model Provider
Here's how to initialize the Model Provider:
import ballerina/ai; import ballerinax/ai.ollama; final ai:ModelProvider ollamaModel = check new ollama:ModelProvider("ollamaModelName");
Step 4: Invoke chat completion
ai:ChatMessage[] chatMessages = [{role: "user", content: "hi"}]; ai:ChatAssistantMessage response = check ollamaModel->chat(chatMessages, tools = []); chatMessages.push(response);