File tree Expand file tree Collapse file tree 1 file changed +3
-0
lines changed
Expand file tree Collapse file tree 1 file changed +3
-0
lines changed Original file line number Diff line number Diff line change @@ -204,6 +204,7 @@ static ggml_cuda_device_info ggml_cuda_init() {
204204 GGML_LOG_INFO (" %s: GGML_CUDA_FORCE_CUBLAS: no\n " , __func__);
205205#endif // GGML_CUDA_FORCE_CUBLAS
206206 GGML_LOG_INFO (" %s: found %d " GGML_CUDA_NAME " devices:\n " , __func__, info.device_count );
207+
207208 std::vector<std::pair<int , std::string>> affected_devices;
208209 for (int id = 0 ; id < info.device_count ; ++id) {
209210 int device_vmm = 0 ;
@@ -274,6 +275,7 @@ static ggml_cuda_device_info ggml_cuda_init() {
274275 }
275276#endif // defined(GGML_USE_HIP)
276277 }
278+
277279 if (ggml_cuda_highest_compiled_arch (GGML_CUDA_CC_TURING) >= GGML_CUDA_CC_TURING && affected_devices.size ()) {
278280 GGML_LOG_INFO (" These devices will have affected performance due to lack of tensor cores:\n " );
279281 for (size_t affected_id = 0 ; affected_id < affected_devices.size (); affected_id++) {
@@ -284,6 +286,7 @@ static ggml_cuda_device_info ggml_cuda_init() {
284286 " If you don't plan to use other devices, consider compiling with different options (e.g. "
285287 " `-DCMAKE_CUDA_ARCHITECTURES=61 -DGGML_CUDA_FORCE_MMQ=1`) for better performance.\n " );
286288 }
289+
287290 for (int id = 0 ; id < info.device_count ; ++id) {
288291 info.default_tensor_split [id] /= total_vram;
289292 }
You can’t perform that action at this time.
0 commit comments