diff --git a/LLama.KernelMemory/LLamaSharp.KernelMemory.csproj b/LLama.KernelMemory/LLamaSharp.KernelMemory.csproj index 7dea57b1..8c7d2c8a 100644 --- a/LLama.KernelMemory/LLamaSharp.KernelMemory.csproj +++ b/LLama.KernelMemory/LLamaSharp.KernelMemory.csproj @@ -4,7 +4,7 @@ net8.0 enable enable - 0.19.0 + 0.20.0 Xbotter SciSharp STACK true diff --git a/LLama.SemanticKernel/LLamaSharp.SemanticKernel.csproj b/LLama.SemanticKernel/LLamaSharp.SemanticKernel.csproj index 674f5761..f332bcf7 100644 --- a/LLama.SemanticKernel/LLamaSharp.SemanticKernel.csproj +++ b/LLama.SemanticKernel/LLamaSharp.SemanticKernel.csproj @@ -10,7 +10,7 @@ enable enable - 0.19.0 + 0.20.0 Tim Miller, Xbotter SciSharp STACK true diff --git a/LLama/LLamaSharp.csproj b/LLama/LLamaSharp.csproj index a32aa764..7944b79d 100644 --- a/LLama/LLamaSharp.csproj +++ b/LLama/LLamaSharp.csproj @@ -7,7 +7,7 @@ AnyCPU;x64;Arm64 True - 0.19.0 + 0.20.0 Rinne, Martin Evans, jlsantiago and all the other contributors in https://github.com/SciSharp/LLamaSharp/graphs/contributors. SciSharp STACK true @@ -22,7 +22,7 @@ With the higher-level APIs and RAG support, it's convenient to deploy LLM (Large Language Model) in your application with LLamaSharp. - Updated llama.cpp version to c35e586ea57221844442c65a1172498c54971cb0 + Updated llama.cpp version to 0827b2c1da299805288abbd556d869318f2b121e MIT packages diff --git a/README.md b/README.md index 9a76cedc..0b556190 100644 --- a/README.md +++ b/README.md @@ -260,6 +260,7 @@ If you want to compile llama.cpp yourself you **must** use the exact commit ID l | v0.17.0 | | [`c35e586e`](https://github.com/ggerganov/llama.cpp/tree/c35e586ea57221844442c65a1172498c54971cb0) | | v0.18.0 | | [`c35e586e`](https://github.com/ggerganov/llama.cpp/tree/c35e586ea57221844442c65a1172498c54971cb0) | | v0.19.0 | | [`958367bf`](https://github.com/ggerganov/llama.cpp/tree/958367bf530d943a902afa1ce1c342476098576b) | +| v0.20.0 | | [`0827b2c1`](https://github.com/ggerganov/llama.cpp/tree/0827b2c1da299805288abbd556d869318f2b121e) | ## License