We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent e050cdf commit c24c6ffCopy full SHA for c24c6ff
llama/addon/globals/getMemoryInfo.cpp
@@ -48,9 +48,9 @@ Napi::Value getMemoryInfo(const Napi::CallbackInfo& info) {
48
addonLlamaCppLogCallback(GGML_LOG_LEVEL_ERROR, std::string("Failed to get memory usage info").c_str(), nullptr);
49
}
50
#elif _WIN32
51
- PROCESS_MEMORY_COUNTERS memCounters;
52
-
53
- if (GetProcessMemoryInfo(GetCurrentProcess(), &memCounters, sizeof(memCounters))) {
+ PROCESS_MEMORY_COUNTERS_EX memCounters;
+
+ if (GetProcessMemoryInfo(GetCurrentProcess(), (PROCESS_MEMORY_COUNTERS*)&memCounters, sizeof(memCounters))) {
54
totalMemoryUsage = memCounters.PrivateUsage;
55
} else {
56
0 commit comments