Releases: SciSharp/LLamaSharp
Releases · SciSharp/LLamaSharp
v0.22.0
Major Changes
- Grammar Resampling by @martindevans, @m0nsky & @phil-scott-78 in #1109
- Added 'strict' parameter for template retrieval by @Lyrcaxis in #1124
- March 2025 version update by @martindevans in #1126
Bug Fixes
- Various fixes regarding special tokens (embeddings & anti-prompt) by @Lyrcaxis in #1112
- Fix Kernel Memory by @martindevans in #1081
- Updates json gbnf to match llama.cpp example by @phil-scott-78 in #1098
- InteractiveExecutor now stops at EOS tokens again by @dpmm99 in #1092
Other Changes
- Update README.md by @sangyuxiaowu in #1086
- Smaller Unit Test Model by @martindevans in #1105
- Update Microsoft.Extensions.AI to 9.3.0-preview.1.25114.11 by @stephentoub in #1102
- Cleaned up non-important warnings from the error list by @Lyrcaxis in #1111
- Update to M.E.AI 9.3.0-preview.1.25161.3 by @stephentoub in #1127
- Improved Memory Rental by @martindevans in #1113
Full Changelog: v0.21.0...v0.22.0
v0.21.0
v0.20.0
Major Changes
- January 2025 Update by @m0nsky, @Lyrcaxis, @SignalRT & @martindevans in #1036
- Removed .NET6 by @martindevans in #982
- Implemented
LogitBias
forDefaultSamplingPipeline
by @martindevans in #986
Bug Fixes
- Fixed
LLamaTokenDataArrayNative.Create
's size initialization by @Lyrcaxis in #981 - Fixed
ChatSession.LoadSession
by @martindevans in #976 - Non-deterministic default seed by @Lyrcaxis in #993
- Fixed log levels (issue #995) by @LoicDagnas in #997
- Fix Batching shared token issues by @martindevans in #1045
Other Changes
- Adding simple batch example by @phil-scott-78 in #1038
- Add an option to apply the chat template to prompts when using StateslessExecutor by @phil-scott-78 in #1025
- Adds an exception for missing templates. by @phil-scott-78 in #1034
- Updated tests model to "LLama 3.2 1B Instruct" by @Lyrcaxis in #983
- Update M.E.AI.Abstractions by @stephentoub in #979, #992 and #1044
- Added sampling tests by @Lyrcaxis in #984
- Updated Nuget Dependencies by @martindevans in #987
- Helper method for sampling
Conversation
withSafeLLamaSamplerChainHandle
by @martindevans in #994
New Contributors
- @LoicDagnas made their first contribution in #997
- @AcoranGonzalezMoray made their first contribution in #1015
- @phil-scott-78 made their first contribution in #1025
Full Changelog: v0.19.0...v0.20.0
v0.19.0
Major Changes
- November Binary Update by @martindevans in #962
- Custom Sampler Stages by @martindevans in #961
- Add Microsoft.Extensions.AI support for IChatClient / IEmbeddingGenerator by @stephentoub in #964
Bug Fixes
- Non-Null Default
SamplingPipeline
by @martindevans in #973
New Contributors
- @stephentoub made their first contribution in #964
Full Changelog: v0.18.0...v0.19.0
v0.18.0
Major Changes
Other Changes
- Updates to ContributingGuide for latest llama.cpp repo by @scritch1sm in #953
- Fix README chat session example by @easis in #956
New Contributors
Full Changelog: v0.17.0...v0.18.0
v0.17.0
Important: The CUDA packages for 0.17.0 exceeded the maximum size for a nuget package. This means some of the 0.17.0 packages are not available until a new way is to deploy those packages. If you need one of the missing packages, use 0.16.0 instead.
Major Changes
- October binary update by @martindevans in #940
Bug Fixes
- Fix: Increase Vulkan detection timeout to 12 seconds by @LSXPrime in #931
- Specify file open permissions needed as read only. by @scritch1sm in #943
- Fix
llama_get_logits_ith
Null Handling by @martindevans in #920 - Fix CUDA compile error by @martindevans in #924
Other Changes
- Add LLama2 Chat Session example with a custom templator by @asmirnov82 in #938
New Contributors
- @LSXPrime made their first contribution in #931
- @scritch1sm made their first contribution in #943
Full Changelog: v0.16.0...v0.17.0
v0.16.0
Major Changes
- August 2024 Binary Update by @martindevans in #905
- LLamaEmbedder 2.0 by @martindevans in #902
Bug Fixes
- Fixed constrained output (Grammar) support in BatchedExecutorGuidance by @webitube in #904
- Fixed Issue where Model with Grammar constrains only generates one output in Example by @L-Wehmschulte in #908
Other Changes
- Remove OpenCL & add Vulkan to readme by @m0nsky in #907
- Bump Swashbuckle.AspNetCore from 6.6.2 to 6.7.3 by @dependabot in #912
- Bump BenchmarkDotNet from 0.13.12 to 0.14.0 by @dependabot in #913
- Bump Microsoft.NET.Test.Sdk from 17.10.0 to 17.11.0 by @dependabot in #914
- Bump Microsoft.AspNetCore.Mvc.Razor.RuntimeCompilation from 8.0.7 to 8.0.8 by @dependabot in #916
- Bump Microsoft.KernelMemory.Abstractions and Microsoft.SemanticKernel.Abstractions by @dependabot in #915
New Contributors
- @webitube made their first contribution in #904
- @L-Wehmschulte made their first contribution in #908
Full Changelog: v0.15.0...v0.16.0
v0.15.0 - Llama 3.1
Major Changes
- Llama 3.1 update binaries by @martindevans in #874
- Microsoft.KernelMemory version 0.68+ compatibility fix by @SpaceAntelope in #862
Bug Fixes
- Fixed missing
[DllImport]
forllama_encode
by @martindevans in #864 - Safer Vulkan System Info by @martindevans in #885
- Fixed LLamaTimings by @martindevans in #881
Other Changes
- Fixed Incorrect Exception Docs Comments by @martindevans in #883
- Deprecate
InferenceParams
sampler configuration properties by @martindevans in #871 - Updated
Semantic Kernel
andKernel Memory
by @martindevans in #880 - Update README.md by @sangyuxiaowu in #872
New Contributors
- @SpaceAntelope made their first contribution in #862
Full Changelog: v0.14.0...v0.15.0
0.14.0 - Gemma2
Major Changes
- July 2024 binaries by @martindevans, @SignalRT & @m0nsky in #833, #843
- Vulkan support by @m0nsky in #794, #797, #799
- Generic Prompt Formatter by @patrick-hovsepian in #787
- Async implementation of LLamaExecutors by @asmirnov82 in #834
- Include debug symbols with NuGet packages by @dpmm99 in #851
- Allow continuation in Instruct and Interact executors by @dpmm99 in #852
Bug Fixes
- Fixes Typo in IContextParamsExtensions by @SerialKicked in #806
- Fix possible file locks due to race conditions during build & test by @m0nsky in #836
- Fixed CI Out Of Disk Space by @martindevans in #846
- Fix CUDA detection on WSL by @m0nsky in #847
Other Changes
- Remove Beam Search by @martindevans in #783
- Batched Beam Search by @martindevans in #796
- Remove OpenCL Backend. by @martindevans in #782
- BoolQ Benchmark by @martindevans in #802
New Contributors
- @patrick-hovsepian made their first contribution in #787
- @SerialKicked made their first contribution in #806
- @dpmm99 made their first contribution in #851
Full Changelog: v0.13.0...v0.14.0
0.13.0
Major Changes
- June 2024 Binary Update by @martindevans in #751
- Add Rosetta2 Binaries by @abhiaagarwal in #755
- Optimize the native library loading. by @AsakusaRinne in #688
- Add experimental auto-download support. by @AsakusaRinne in #692
- Upload native libraries to huggingface when release new version. by @AsakusaRinne in #749
- Batch Queueing by @martindevans in #761
- BatchedExecutor LLava by @martindevans in #770
- StreamingTextDecoder Special Token Support by @martindevans in #777
Documentation
- Updated READMEs. by @Lamothe in #742
- Add documentation for the change of native library loading. by @AsakusaRinne in #747
- Add OAI demo link by @sangyuxiaowu in #778
Bug Fixes
- KernelMemory bug fix by @zsogitbe in #726
- Fixes for website when running on case-sensitive filesystems. by @Lamothe in #753
Other Changes
- Exposed basic timing information from llama.cpp by @martindevans in #736
- Less Sampler Allocations by @martindevans in #735
- Several updates to web project by @Lamothe in #718
- Conversation Generate All Logits by @martindevans in #743
- Removed .NET 7 now that it's reached EOL. by @Lamothe in #741
- BatchedExecutor Double Buffering by @martindevans in #748
- Minor BatchedExecutor Example Improvements by @martindevans in #750
- LLava API Improvements by @martindevans in #757
- SemanticKernel HomeAutomation Example by @zsogitbe in #769
New Contributors
- @Lamothe made their first contribution in #718
- @sangyuxiaowu made their first contribution in #778
- @abhiaagarwal made their first contribution in #755
Full Changelog: v0.12.0...v0.13.0