Skip to content

Commit 20a5ecd

Browse files
committed
Add newlines
1 parent e236d2e commit 20a5ecd

File tree

1 file changed

+3
-0
lines changed

1 file changed

+3
-0
lines changed

ggml/src/ggml-cuda/ggml-cuda.cu

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -204,6 +204,7 @@ static ggml_cuda_device_info ggml_cuda_init() {
204204
GGML_LOG_INFO("%s: GGML_CUDA_FORCE_CUBLAS: no\n", __func__);
205205
#endif // GGML_CUDA_FORCE_CUBLAS
206206
GGML_LOG_INFO("%s: found %d " GGML_CUDA_NAME " devices:\n", __func__, info.device_count);
207+
207208
std::vector<std::pair<int, std::string>> affected_devices;
208209
for (int id = 0; id < info.device_count; ++id) {
209210
int device_vmm = 0;
@@ -274,6 +275,7 @@ static ggml_cuda_device_info ggml_cuda_init() {
274275
}
275276
#endif // defined(GGML_USE_HIP)
276277
}
278+
277279
if (ggml_cuda_highest_compiled_arch(GGML_CUDA_CC_TURING) >= GGML_CUDA_CC_TURING && affected_devices.size()) {
278280
GGML_LOG_INFO("These devices will have affected performance due to lack of tensor cores:\n");
279281
for (size_t affected_id = 0; affected_id < affected_devices.size(); affected_id++) {
@@ -284,6 +286,7 @@ static ggml_cuda_device_info ggml_cuda_init() {
284286
"If you don't plan to use other devices, consider compiling with different options (e.g. "
285287
"`-DCMAKE_CUDA_ARCHITECTURES=61 -DGGML_CUDA_FORCE_MMQ=1`) for better performance.\n");
286288
}
289+
287290
for (int id = 0; id < info.device_count; ++id) {
288291
info.default_tensor_split[id] /= total_vram;
289292
}

0 commit comments

Comments
 (0)