From 06c12b627c4b2da94c05b021a6b2e31539557798 Mon Sep 17 00:00:00 2001 From: Adrian Cole <64215+codefromthecrypt@users.noreply.github.com> Date: Tue, 15 Oct 2024 14:09:03 +1100 Subject: [PATCH] chore: updates ollama default model from mistral-nemo to qwen2.5 (#150) Signed-off-by: Adrian Cole --- packages/exchange/src/exchange/providers/ollama.py | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/packages/exchange/src/exchange/providers/ollama.py b/packages/exchange/src/exchange/providers/ollama.py index 88856464..f05ea426 100644 --- a/packages/exchange/src/exchange/providers/ollama.py +++ b/packages/exchange/src/exchange/providers/ollama.py @@ -6,24 +6,24 @@ import httpx from exchange.providers.openai import OpenAiProvider OLLAMA_HOST = "http://localhost:11434/" -OLLAMA_MODEL = "mistral-nemo" +OLLAMA_MODEL = "qwen2.5" class OllamaProvider(OpenAiProvider): """Provides chat completions for models hosted by Ollama.""" - __doc__ += """Here's an example profile configuration to try: + __doc__ += f"""Here's an example profile configuration to try: First run: ollama pull qwen2.5, then use this profile: ollama: provider: ollama - processor: qwen2.5 - accelerator: qwen2.5 + processor: {OLLAMA_MODEL} + accelerator: {OLLAMA_MODEL} moderator: truncate toolkits: - name: developer - requires: {} + requires: {{}} """ def __init__(self, client: httpx.Client) -> None: