diff --git a/LLama.KernelMemory/LLamaSharp.KernelMemory.csproj b/LLama.KernelMemory/LLamaSharp.KernelMemory.csproj
index 7dea57b1..8c7d2c8a 100644
--- a/LLama.KernelMemory/LLamaSharp.KernelMemory.csproj
+++ b/LLama.KernelMemory/LLamaSharp.KernelMemory.csproj
@@ -4,7 +4,7 @@
net8.0
enable
enable
- 0.19.0
+ 0.20.0
Xbotter
SciSharp STACK
true
diff --git a/LLama.SemanticKernel/LLamaSharp.SemanticKernel.csproj b/LLama.SemanticKernel/LLamaSharp.SemanticKernel.csproj
index 674f5761..f332bcf7 100644
--- a/LLama.SemanticKernel/LLamaSharp.SemanticKernel.csproj
+++ b/LLama.SemanticKernel/LLamaSharp.SemanticKernel.csproj
@@ -10,7 +10,7 @@
enable
enable
- 0.19.0
+ 0.20.0
Tim Miller, Xbotter
SciSharp STACK
true
diff --git a/LLama/LLamaSharp.csproj b/LLama/LLamaSharp.csproj
index a32aa764..7944b79d 100644
--- a/LLama/LLamaSharp.csproj
+++ b/LLama/LLamaSharp.csproj
@@ -7,7 +7,7 @@
AnyCPU;x64;Arm64
True
- 0.19.0
+ 0.20.0
Rinne, Martin Evans, jlsantiago and all the other contributors in https://github.com/SciSharp/LLamaSharp/graphs/contributors.
SciSharp STACK
true
@@ -22,7 +22,7 @@
With the higher-level APIs and RAG support, it's convenient to deploy LLM (Large Language Model) in your application with LLamaSharp.
- Updated llama.cpp version to c35e586ea57221844442c65a1172498c54971cb0
+ Updated llama.cpp version to 0827b2c1da299805288abbd556d869318f2b121e
MIT
packages
diff --git a/README.md b/README.md
index 9a76cedc..0b556190 100644
--- a/README.md
+++ b/README.md
@@ -260,6 +260,7 @@ If you want to compile llama.cpp yourself you **must** use the exact commit ID l
| v0.17.0 | | [`c35e586e`](https://github.com/ggerganov/llama.cpp/tree/c35e586ea57221844442c65a1172498c54971cb0) |
| v0.18.0 | | [`c35e586e`](https://github.com/ggerganov/llama.cpp/tree/c35e586ea57221844442c65a1172498c54971cb0) |
| v0.19.0 | | [`958367bf`](https://github.com/ggerganov/llama.cpp/tree/958367bf530d943a902afa1ce1c342476098576b) |
+| v0.20.0 | | [`0827b2c1`](https://github.com/ggerganov/llama.cpp/tree/0827b2c1da299805288abbd556d869318f2b121e) |
## License