@@ -398,6 +398,7 @@ struct ggml_backend_opencl_context {
398398 cl_kernel kernel_scale;
399399 cl_kernel kernel_silu, kernel_silu_4;
400400 cl_kernel kernel_gelu, kernel_gelu_4;
401+ cl_kernel kernel_gelu_erf, kernel_gelu_erf_4;
401402 cl_kernel kernel_gelu_quick, kernel_gelu_quick_4;
402403 cl_kernel kernel_relu;
403404 cl_kernel kernel_sigmoid_f32, kernel_sigmoid_f16;
@@ -736,6 +737,8 @@ static void load_cl_kernels(ggml_backend_opencl_context *backend_ctx, ggml_cl_ve
736737
737738 CL_CHECK ((backend_ctx->kernel_gelu = clCreateKernel (backend_ctx->program_gelu , " kernel_gelu" , &err), err));
738739 CL_CHECK ((backend_ctx->kernel_gelu_4 = clCreateKernel (backend_ctx->program_gelu , " kernel_gelu_4" , &err), err));
740+ CL_CHECK ((backend_ctx->kernel_gelu_erf = clCreateKernel (backend_ctx->program_gelu , " kernel_gelu_erf" , &err), err));
741+ CL_CHECK ((backend_ctx->kernel_gelu_erf_4 = clCreateKernel (backend_ctx->program_gelu , " kernel_gelu_erf_4" , &err), err));
739742 CL_CHECK ((backend_ctx->kernel_gelu_quick = clCreateKernel (backend_ctx->program_gelu , " kernel_gelu_quick" , &err), err));
740743 CL_CHECK ((backend_ctx->kernel_gelu_quick_4 = clCreateKernel (backend_ctx->program_gelu , " kernel_gelu_quick_4" , &err), err));
741744 GGML_LOG_CONT (" ." );
@@ -2256,6 +2259,7 @@ static bool ggml_opencl_supports_op(ggml_backend_dev_t dev, const struct ggml_te
22562259 case GGML_UNARY_OP_GELU:
22572260 case GGML_UNARY_OP_SILU:
22582261 case GGML_UNARY_OP_RELU:
2262+ case GGML_UNARY_OP_GELU_ERF:
22592263 case GGML_UNARY_OP_GELU_QUICK:
22602264 return ggml_is_contiguous (op->src [0 ]) && op->src [0 ]->type == GGML_TYPE_F32;
22612265 case GGML_UNARY_OP_SIGMOID:
@@ -3858,6 +3862,44 @@ static void ggml_cl_gelu(ggml_backend_t backend, const ggml_tensor * src0, const
38583862 backend_ctx->enqueue_ndrange_kernel (kernel, 3 , global_work_size, local_work_size, dst);
38593863}
38603864
3865+ static void ggml_cl_gelu_erf (ggml_backend_t backend, const ggml_tensor * src0, const ggml_tensor * src1, ggml_tensor * dst) {
3866+ GGML_ASSERT (src0);
3867+ GGML_ASSERT (src0->extra );
3868+ GGML_ASSERT (dst);
3869+ GGML_ASSERT (dst->extra );
3870+
3871+ UNUSED (src1);
3872+
3873+ ggml_backend_opencl_context *backend_ctx = (ggml_backend_opencl_context *)backend->context ;
3874+
3875+ ggml_tensor_extra_cl * extra0 = (ggml_tensor_extra_cl *)src0->extra ;
3876+ ggml_tensor_extra_cl * extrad = (ggml_tensor_extra_cl *)dst->extra ;
3877+
3878+ cl_ulong offset0 = extra0->offset + src0->view_offs ;
3879+ cl_ulong offsetd = extrad->offset + dst->view_offs ;
3880+
3881+ cl_kernel kernel;
3882+
3883+ int n = ggml_nelements (dst);
3884+
3885+ if (n % 4 == 0 ) {
3886+ kernel = backend_ctx->kernel_gelu_erf_4 ;
3887+ n /= 4 ;
3888+ } else {
3889+ kernel = backend_ctx->kernel_gelu_erf ;
3890+ }
3891+
3892+ CL_CHECK (clSetKernelArg (kernel, 0 , sizeof (cl_mem), &extra0->data_device ));
3893+ CL_CHECK (clSetKernelArg (kernel, 1 , sizeof (cl_ulong), &offset0));
3894+ CL_CHECK (clSetKernelArg (kernel, 2 , sizeof (cl_mem), &extrad->data_device ));
3895+ CL_CHECK (clSetKernelArg (kernel, 3 , sizeof (cl_ulong), &offsetd));
3896+
3897+ size_t global_work_size[] = {(size_t )n, 1 , 1 };
3898+ size_t local_work_size[] = {64 , 1 , 1 };
3899+
3900+ backend_ctx->enqueue_ndrange_kernel (kernel, 3 , global_work_size, local_work_size, dst);
3901+ }
3902+
38613903static void ggml_cl_gelu_quick (ggml_backend_t backend, const ggml_tensor * src0, const ggml_tensor * src1, ggml_tensor * dst) {
38623904 GGML_ASSERT (src0);
38633905 GGML_ASSERT (src0->extra );
@@ -6329,6 +6371,12 @@ bool ggml_cl_compute_forward(ggml_backend_t backend, struct ggml_tensor * tensor
63296371 }
63306372 func = ggml_cl_gelu;
63316373 break ;
6374+ case GGML_UNARY_OP_GELU_ERF:
6375+ if (!any_on_device) {
6376+ return false ;
6377+ }
6378+ func = ggml_cl_gelu_erf;
6379+ break ;
63326380 case GGML_UNARY_OP_GELU_QUICK:
63336381 if (!any_on_device) {
63346382 return false ;
0 commit comments