File tree Expand file tree Collapse file tree 1 file changed +0
-16
lines changed Expand file tree Collapse file tree 1 file changed +0
-16
lines changed Original file line number Diff line number Diff line change @@ -262,22 +262,6 @@ static bool fp16_mma_hardware_available(const int cc) {
262262        (GGML_CUDA_CC_IS_MTHREADS (cc) && cc >= GGML_CUDA_CC_QY2);
263263}
264264
265- static  bool  bf16_mma_hardware_available (const  int  cc) {
266-     return  (GGML_CUDA_CC_IS_NVIDIA (cc) && cc >= GGML_CUDA_CC_AMPERE) || GGML_CUDA_CC_IS_CDNA (cc) || cc >= GGML_CUDA_CC_RDNA3;
267- }
268- 
269- static  bool  fp32_mma_hardware_available (const  int  cc) {
270-     return  GGML_CUDA_CC_IS_CDNA (cc);
271- }
272- 
273- static  bool  bf16_mma_hardware_available (const  int  cc) {
274-     return  (GGML_CUDA_CC_IS_NVIDIA (cc) && cc >= GGML_CUDA_CC_AMPERE) || GGML_CUDA_CC_IS_CDNA (cc) || cc >= GGML_CUDA_CC_RDNA3;
275- }
276- 
277- static  bool  fp32_mma_hardware_available (const  int  cc) {
278-     return  GGML_CUDA_CC_IS_CDNA (cc);
279- }
280- 
281265//  Volta technically had FP16 tensor cores but they work very differently compared to Turing and later.
282266static  bool  new_mma_available (const  int  cc) {
283267    return  GGML_CUDA_CC_IS_NVIDIA (cc) && ggml_cuda_highest_compiled_arch (cc) >= GGML_CUDA_CC_TURING;
    
 
   
 
     
   
   
          
     
  
    
     
 
    
      
     
 
     
    You can’t perform that action at this time.
  
 
    
  
     
    
      
        
     
 
       
      
     
   
 
    
    
  
 
  
 
     
    
0 commit comments