Skip to content

Commit 4341e83

Browse files
committed
Updated to 11dd5a44eb180e1d69fac24d3852b5222d66fb7f
1 parent 992aec8 commit 4341e83

File tree

5 files changed

+16
-2
lines changed

5 files changed

+16
-2
lines changed

LLama.Web/Common/ModelOptions.cs

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -116,6 +116,9 @@ public class ModelOptions
116116
/// <inheritdoc />
117117
public bool? SwaFull { get; set; }
118118

119+
/// <inheritdoc />
120+
public bool? KVUnified { get; set; }
121+
119122
/// <inheritdoc />
120123
public float? DefragThreshold { get; set; }
121124

LLama/Abstractions/IContextParams.cs

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -128,6 +128,14 @@ public interface IContextParams
128128
/// </summary>
129129
bool? OpOffload { get; }
130130

131+
/// <summary>
132+
/// use a unified buffer across the input sequences when computing the attention.
133+
/// try to disable when n_seq_max > 1 for improved performance when the sequences do not share a large prefix
134+
/// <br />
135+
/// ref: <a href="https://github.com/ggml-org/llama.cpp/pull/14363">https://github.com/ggml-org/llama.cpp/pull/14363</a>
136+
/// </summary>
137+
bool? KVUnified { get; }
138+
131139
/// <summary>
132140
/// Use full-size SWA cache (https://github.com/ggml-org/llama.cpp/pull/13194#issuecomment-2868343055)
133141
/// </summary>

LLama/Common/ModelParams.cs

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -118,6 +118,9 @@ public record ModelParams
118118
/// <inheritdoc />
119119
public bool? SwaFull { get; set; }
120120

121+
/// <inheritdoc />
122+
public bool? KVUnified { get; set; }
123+
121124
/// <summary>
122125
/// `Encoding` cannot be directly JSON serialized, instead store the name as a string which can
123126
/// </summary>

LLama/LLamaSharp.csproj

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -57,7 +57,7 @@
5757
</ItemGroup>
5858

5959
<PropertyGroup>
60-
<BinaryReleaseId>a0374a67e2924f2e</BinaryReleaseId>
60+
<BinaryReleaseId>11dd5a44eb180e</BinaryReleaseId>
6161
</PropertyGroup>
6262

6363
<PropertyGroup>

llama.cpp

0 commit comments

Comments
 (0)