From 2bc1f1ef09b56e24d128b20ee711415c487c9bf9 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 9 Feb 2026 20:42:00 +0000 Subject: [PATCH 1/2] build(deps): bump llama-cpp-2 from 0.1.132 to 0.1.133 Bumps [llama-cpp-2](https://github.com/utilityai/llama-cpp-rs) from 0.1.132 to 0.1.133. - [Release notes](https://github.com/utilityai/llama-cpp-rs/releases) - [Commits](https://github.com/utilityai/llama-cpp-rs/compare/0.1.132...0.1.133) --- updated-dependencies: - dependency-name: llama-cpp-2 dependency-version: 0.1.133 dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] --- Cargo.lock | 33 +++++++++++++++++---------------- crates/core/Cargo.toml | 2 +- 2 files changed, 18 insertions(+), 17 deletions(-) diff --git a/Cargo.lock b/Cargo.lock index 8e66591..a0ae481 100644 --- a/Cargo.lock +++ b/Cargo.lock @@ -388,9 +388,9 @@ checksum = "1e748733b7cbc798e1434b6ac524f0c1ff2ab456fe201501e6497c8417a4fc33" [[package]] name = "cc" -version = "1.2.51" +version = "1.2.55" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "7a0aeaff4ff1a90589618835a598e545176939b97874f7abc7851caa0618f203" +checksum = "47b26a0954ae34af09b50f0de26458fa95369a0d478d8236d3f93082b219bd29" dependencies = [ "find-msvc-tools", "jobserver", @@ -682,7 +682,7 @@ dependencies = [ "libc", "option-ext", "redox_users", - "windows-sys 0.59.0", + "windows-sys 0.61.2", ] [[package]] @@ -785,7 +785,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index" checksum = "39cab71617ae0d63f51a36d69f866391735b51691dbda63cf6f96d042b63efeb" dependencies = [ "libc", - "windows-sys 0.59.0", + "windows-sys 0.61.2", ] [[package]] @@ -824,9 +824,9 @@ dependencies = [ [[package]] name = "find-msvc-tools" -version = "0.1.6" +version = "0.1.9" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "645cbb3a84e60b7531617d5ae4e57f7e27308f6445f5abf653209ea76dec8dff" +checksum = "5baebc0774151f905a1a2cc41989300b1e6fbb29aff0ceffa1064fdd3088d582" [[package]] name = "find_cuda_helper" @@ -1521,10 +1521,11 @@ checksum = "6373607a59f0be73a39b6fe456b8192fcc3585f602af20751600e974dd455e77" [[package]] name = "llama-cpp-2" -version = "0.1.132" +version = "0.1.133" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "ce13ae10879183e15e093f514d52bf128901fc489a989969512526a7063088cf" +checksum = "888c8805527f4c35ec16f26003d54a318cde1629e7439da8e9ef2d6d3883e106" dependencies = [ + "encoding_rs", "enumflags2", "llama-cpp-sys-2", "thiserror 1.0.69", @@ -1534,9 +1535,9 @@ dependencies = [ [[package]] name = "llama-cpp-sys-2" -version = "0.1.132" +version = "0.1.133" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "2fcb116a1a687f928484bcd257f811e6e0665dcb39b53a122725eca12324cf9e" +checksum = "a180dfa6d6f9d1df1e031bcdf0464bbad4f9b326395bfd28f2fa539d8cbc9c2b" dependencies = [ "bindgen", "cc", @@ -1716,7 +1717,7 @@ version = "0.50.3" source = "registry+https://github.com/rust-lang/crates.io-index" checksum = "7957b9740744892f114936ab4a57b3f487491bbeafaf8083688b16841a4240e5" dependencies = [ - "windows-sys 0.59.0", + "windows-sys 0.61.2", ] [[package]] @@ -2006,7 +2007,7 @@ dependencies = [ "once_cell", "socket2", "tracing", - "windows-sys 0.59.0", + "windows-sys 0.60.2", ] [[package]] @@ -2279,7 +2280,7 @@ dependencies = [ "errno", "libc", "linux-raw-sys", - "windows-sys 0.59.0", + "windows-sys 0.61.2", ] [[package]] @@ -2337,7 +2338,7 @@ dependencies = [ "security-framework", "security-framework-sys", "webpki-root-certs", - "windows-sys 0.59.0", + "windows-sys 0.61.2", ] [[package]] @@ -2691,7 +2692,7 @@ dependencies = [ "getrandom 0.3.4", "once_cell", "rustix", - "windows-sys 0.59.0", + "windows-sys 0.61.2", ] [[package]] @@ -3208,7 +3209,7 @@ version = "0.1.11" source = "registry+https://github.com/rust-lang/crates.io-index" checksum = "c2a7b1c03c876122aa43f3020e6c3c3ee5c05081c9a00739faf7503aeba10d22" dependencies = [ - "windows-sys 0.59.0", + "windows-sys 0.61.2", ] [[package]] diff --git a/crates/core/Cargo.toml b/crates/core/Cargo.toml index 67fd6ad..4df2124 100644 --- a/crates/core/Cargo.toml +++ b/crates/core/Cargo.toml @@ -36,7 +36,7 @@ tracing-subscriber.workspace = true notify.workspace = true minijinja = { version = "2.15.1", features = ["builtins", "json"] } minijinja-contrib = { version = "2.15.1", features = ["pycompat"] } -llama-cpp-2 = "=0.1.132" +llama-cpp-2 = "=0.1.133" [dev-dependencies] tempfile = { workspace = true, features = [] } From ab3829866c5e735e68b99dc3868d629bfaf0a0ca Mon Sep 17 00:00:00 2001 From: Arun Mahapatra Date: Fri, 13 Feb 2026 04:49:51 +0530 Subject: [PATCH 2/2] fix: replace deprecated methods --- crates/core/src/provider/gguf/mod.rs | 16 +++++++++++----- 1 file changed, 11 insertions(+), 5 deletions(-) diff --git a/crates/core/src/provider/gguf/mod.rs b/crates/core/src/provider/gguf/mod.rs index 0f33114..8fb4a5b 100644 --- a/crates/core/src/provider/gguf/mod.rs +++ b/crates/core/src/provider/gguf/mod.rs @@ -8,12 +8,11 @@ use anyhow::{Context, Result, anyhow}; use async_trait::async_trait; use futures::stream::BoxStream; use llama_cpp_2::token::LlamaToken; -use llama_cpp_2::{LogOptions, send_logs_to_tracing}; +use llama_cpp_2::{LogOptions, TokenToStringError, send_logs_to_tracing}; use llama_cpp_2::{ context::params::LlamaContextParams, llama_backend::LlamaBackend, llama_batch::LlamaBatch, - model::Special, model::{AddBos, LlamaModel, params::LlamaModelParams}, sampling::LlamaSampler, }; @@ -293,9 +292,16 @@ impl CompletionModel for GgufBaseModel { } // Get token bytes and decode - let token_bytes = model - .token_to_bytes(token, Special::Tokenize) - .map_err(|e| anyhow!("Token conversion failed: {e}"))?; + let token_bytes = match model.token_to_piece_bytes(token, 8, true, None) { + Err(TokenToStringError::InsufficientBufferSpace(required)) => { + let required = (-required) + .try_into() + .expect("Error buffer size is positive"); + model.token_to_piece_bytes(token, required, true, None) + } + res => res, + } + .map_err(|e| anyhow!("Token conversion failed: {e}"))?; let mut last_chunk = String::with_capacity(32); let _ = decoder.decode_to_string(&token_bytes, &mut last_chunk, false);