From 8f626c0f188727f046f60d5e5f3af6547d63b1ac Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Sergio=20G=C3=B3mez?= Date: Tue, 27 Aug 2024 18:08:39 +0200 Subject: [PATCH] chore: version 0.13.0 --- packages/ollama/CHANGELOG.md | 8 +++++++- packages/ollama/package.json | 2 +- 2 files changed, 8 insertions(+), 2 deletions(-) diff --git a/packages/ollama/CHANGELOG.md b/packages/ollama/CHANGELOG.md index 7f719a6..62faa04 100644 --- a/packages/ollama/CHANGELOG.md +++ b/packages/ollama/CHANGELOG.md @@ -1,5 +1,11 @@ # ollama-ai-provider +## 0.13.0 + +### Minor Changes + +- Add custom provider support + ## 0.12.1 ### Patch Changes @@ -13,7 +19,7 @@ - Update to Vercel AI SDK 3.3 - Added support to sending custom headers - Added support to stopSequences and topK settings -- Added settings to disable experimental stream tooling +- Added settings to disable experimental stream tooling ## 0.11.0 diff --git a/packages/ollama/package.json b/packages/ollama/package.json index 283e87f..f072522 100644 --- a/packages/ollama/package.json +++ b/packages/ollama/package.json @@ -1,6 +1,6 @@ { "name": "ollama-ai-provider", - "version": "0.12.1", + "version": "0.13.0", "description": "Vercel AI Provider for running LLMs locally using Ollama", "main": "./dist/index.js", "module": "./dist/index.mjs",