@@ -270,7 +270,7 @@ void SELayer<float>::Eval(int N, float* output, const float* input,
270
270
cublasHandle_t handle = cuBlasContextManager::getcuBlasHandle_t ();
271
271
// ReportCUBLASErrors(cublasCreate(&handle));
272
272
273
- sycl_queue_.submit ([&](sycl::handler &cgh) {
273
+ sycl_queue_.submit ([&](sycl::handler &cgh) {
274
274
// auto d_A = b_A.get_access<sycl::access::mode::read_write>(cgh);
275
275
276
276
cgh.host_task ([=](sycl::interop_handle ih) {
@@ -417,9 +417,9 @@ void SELayer<sycl::half>::Eval(int N, sycl::half* output, const sycl::half* inpu
417
417
q_ct1->submit([&](sycl::handler &cgh) {
418
418
//auto d_A = b_A.get_access<sycl::access::mode::read_write>(cgh);
419
419
420
- cgh.interop_task ([=](sycl::interop_handler ih) {
420
+ cgh.host_task ([=](sycl::interop_handle ih) {
421
421
422
- auto cudaStreamHandle = sycl::get_native<sycl::backend::cuda >(*q_ct1);
422
+ auto cudaStreamHandle = sycl::get_native<sycl::backend::ext_oneapi_cuda >(*q_ct1);
423
423
cublasSetStream(handle, cudaStreamHandle);
424
424
425
425
ReportCUBLASErrors(cublasHgemm(handle, CUBLAS_OP_T, CUBLAS_OP_N, numFc1Out_,
@@ -439,9 +439,9 @@ void SELayer<sycl::half>::Eval(int N, sycl::half* output, const sycl::half* inpu
439
439
q_ct1->submit([&](sycl::handler &cgh) {
440
440
//auto d_A = b_A.get_access<sycl::access::mode::read_write>(cgh);
441
441
442
- cgh.interop_task ([=](sycl::interop_handler ih) {
442
+ cgh.host_task ([=](sycl::interop_handle ih) {
443
443
444
- auto cudaStreamHandle = sycl::get_native<sycl::backend::cuda >(*q_ct1);
444
+ auto cudaStreamHandle = sycl::get_native<sycl::backend::ext_oneapi_cuda >(*q_ct1);
445
445
cublasSetStream(handle, cudaStreamHandle);
446
446
447
447
// 3. Second fully connected layer.
@@ -556,9 +556,9 @@ void FCLayer<float>::LoadWeights(float* cpuWeight, float* cpuBias,
556
556
557
557
// q_ct1->submit([&](sycl::handler &cgh) {
558
558
559
- // cgh.interop_task ([=](sycl::interop_handler ih) {
559
+ // cgh.host_task ([=](sycl::interop_handle ih) {
560
560
561
- // auto cudaStreamHandle = sycl::get_native<sycl::backend::cuda >(*q_ct1);
561
+ // auto cudaStreamHandle = sycl::get_native<sycl::backend::ext_oneapi_cuda >(*q_ct1);
562
562
// cublasSetStream(handle, cudaStreamHandle);
563
563
564
564
// ReportCUBLASErrors(cublasHgemm(handle, CUBLAS_OP_T, CUBLAS_OP_N, num_outputs,
@@ -918,9 +918,9 @@ void FusedWinogradConvSELayer<DataType>::LoadSEWeights(float* w1, float* b1,
918
918
919
919
// q_ct1->submit([&](sycl::handler &cgh) {
920
920
// //auto d_A = b_A.get_access<sycl::access::mode::read_write>(cgh);
921
- // cgh.interop_task ([=](sycl::interop_handler ih) {
921
+ // cgh.host_task ([=](sycl::interop_handle ih) {
922
922
923
- // auto cudaStreamHandle = sycl::get_native<sycl::backend::cuda >(q_ct1);
923
+ // auto cudaStreamHandle = sycl::get_native<sycl::backend::ext_oneapi_cuda >(q_ct1);
924
924
// cublasSetStream(handle, cudaStreamHandle);
925
925
926
926
// ReportCUBLASErrors(cublasGemmStridedBatchedEx(
@@ -1019,7 +1019,7 @@ template <> void BaseLayer<float>::cublasRowMajorMatrixMul(const float* A, const
1019
1019
else {
1020
1020
1021
1021
#ifdef USE_CUBLAS
1022
- sycl_queue_.submit ([&](sycl::handler &cgh) {
1022
+ sycl_queue_.submit ([&](sycl::handler &cgh) {
1023
1023
// auto d_A = b_A.get_access<sycl::access::mode::read_write>(cgh);
1024
1024
cgh.host_task ([=](sycl::interop_handle ih) {
1025
1025
auto cudaStreamHandle = sycl::get_native<sycl::backend::ext_oneapi_cuda>(sycl_queue_);
@@ -1218,9 +1218,9 @@ template <typename DataType> void Conv1Layer<DataType>::LoadWeights(float* pfilt
1218
1218
1219
1219
// q_ct1->submit([&](sycl::handler &cgh) {
1220
1220
// //auto d_A = b_A.get_access<sycl::access::mode::read_write>(cgh);
1221
- // cgh.interop_task ([=](sycl::interop_handler ih) {
1221
+ // cgh.host_task ([=](sycl::interop_handle ih) {
1222
1222
1223
- // auto cudaStreamHandle = sycl::get_native<sycl::backend::cuda >(q_ct1);
1223
+ // auto cudaStreamHandle = sycl::get_native<sycl::backend::ext_oneapi_cuda >(q_ct1);
1224
1224
// cublasSetStream(handle, cudaStreamHandle);
1225
1225
1226
1226
@@ -1264,7 +1264,7 @@ void Conv1Layer<float>::cublasSpecialMatrixMul(const float* A, const float* B,
1264
1264
1265
1265
// printf("use_gemm_ex_\n");
1266
1266
#ifdef USE_CUBLAS
1267
- sycl_queue_.submit ([&](sycl::handler &cgh) {
1267
+ sycl_queue_.submit ([&](sycl::handler &cgh) {
1268
1268
// auto d_A = b_A.get_access<sycl::access::mode::read_write>(cgh);
1269
1269
cgh.host_task ([=](sycl::interop_handle ih) {
1270
1270
@@ -1282,7 +1282,7 @@ void Conv1Layer<float>::cublasSpecialMatrixMul(const float* A, const float* B,
1282
1282
});
1283
1283
});
1284
1284
#elifdef USE_HIPBLAS
1285
- sycl_queue_.submit ([&](sycl::handler &cgh) {
1285
+ sycl_queue_.submit ([&](sycl::handler &cgh) {
1286
1286
// auto d_A = b_A.get_access<sycl::access::mode::read_write>(cgh);
1287
1287
cgh.host_task ([=](sycl::interop_handle ih) {
1288
1288
@@ -1343,7 +1343,7 @@ void Conv1Layer<float>::cublasSpecialMatrixMul(const float* A, const float* B,
1343
1343
});
1344
1344
});
1345
1345
#elifdef USE_HIPBLAS
1346
- sycl_queue_.submit ([&](sycl::handler &cgh) {
1346
+ sycl_queue_.submit ([&](sycl::handler &cgh) {
1347
1347
// auto d_A = b_A.get_access<sycl::access::mode::read_write>(cgh);
1348
1348
cgh.host_task ([=](sycl::interop_handle ih) {
1349
1349
@@ -1831,9 +1831,9 @@ static void cublasXgemm(transpose_type transa,
1831
1831
1832
1832
q_ct1.submit([&](sycl::handler &cgh) {
1833
1833
1834
- cgh.interop_task ([=](sycl::interop_handler ih) {
1834
+ cgh.host_task ([=](sycl::interop_handle ih) {
1835
1835
1836
- auto cudaStreamHandle = sycl::get_native<sycl::backend::cuda >(q_ct1);
1836
+ auto cudaStreamHandle = sycl::get_native<sycl::backend::ext_oneapi_cuda >(q_ct1);
1837
1837
cublasSetStream(handle, cudaStreamHandle);
1838
1838
1839
1839
@@ -1850,7 +1850,7 @@ static void cublasXgemm(transpose_type transa,
1850
1850
1851
1851
1852
1852
1853
- sycl_queue.submit ([&](sycl::handler &cgh) {
1853
+ sycl_queue.submit ([&](sycl::handler &cgh) {
1854
1854
1855
1855
cgh.host_task ([=](sycl::interop_handle ih) {
1856
1856
@@ -1919,9 +1919,9 @@ static void cublasXGemmStridedBatched(transpose_type transa, transpose_type tran
1919
1919
1920
1920
q_ct1.submit([&](sycl::handler &cgh) {
1921
1921
1922
- cgh.interop_task ([=](sycl::interop_handler ih) {
1922
+ cgh.host_task ([=](sycl::interop_handle ih) {
1923
1923
1924
- auto cudaStreamHandle = sycl::get_native<sycl::backend::cuda >(q_ct1);
1924
+ auto cudaStreamHandle = sycl::get_native<sycl::backend::ext_oneapi_cuda >(q_ct1);
1925
1925
cublasSetStream(handle, cudaStreamHandle);
1926
1926
1927
1927
ReportCUBLASErrors(cublasGemmStridedBatchedEx(
@@ -1956,7 +1956,7 @@ static void cublasXGemmStridedBatched(transpose_type transa, transpose_type tran
1956
1956
1957
1957
1958
1958
#elifdef USE_HIPBLAS
1959
- hipblasHandle_t handle = hipBlasContextManager::gethipBlasHandle_t ();
1959
+ hipblasHandle_t handle = hipBlasContextManager::gethipBlasHandle_t ();
1960
1960
1961
1961
sycl_queue.submit ([&](sycl::handler &cgh) {
1962
1962
0 commit comments