Skip to content

Commit 8c10e5d

Browse files
committed
Update LLamaEmbedder, Examples packages, and KernelMemory examples
- Embedding generation: Extension with Batch processing + Normalization (important to have this built-in for KernelMemory). - Examples had wrong nuget packages, updated to correct ones. - Updated KernelMemory examples. - added missing model parameters - adding config is null check - unit tests project update to prevent the constant download of many GBs - ** for some reason Embeddings must be set to false in the kernel memory text embedding generator => we need to follow this and check it later because this should normally be 'true' ! ** - skipping one test for macOS (all other tests are OK) - setting GpuLayerCount to 0 in Release in CIGpuLayerCount also for Windows - possible BUG in llama.cpp in 'if (params.split_mode == LLAMA_SPLIT_MODE_NONE)'... trying to set other split mode (even if there is no GPU)!
1 parent e9a35cb commit 8c10e5d

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

LLama.Unittest/KernelMemory/ITextTokenizerTests.cs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ public ITextTokenizerTests(ITestOutputHelper testOutputHelper)
2222
_testOutputHelper = testOutputHelper;
2323

2424
_infParams = new() { AntiPrompts = ["\n\n"] };
25-
_lsConfig = new(Constants.GenerativeModelPath) { DefaultInferenceParams = _infParams, ContextSize = 512 };
25+
_lsConfig = new(Constants.GenerativeModelPath) { DefaultInferenceParams = _infParams, ContextSize = 512, SplitMode = LLama.Native.GPUSplitMode.Layer };
2626

2727
testOutputHelper.WriteLine($"Using model {Path.GetFileName(_lsConfig.ModelPath)}");
2828
}

0 commit comments

Comments
 (0)