Skip to content
Merged
Show file tree
Hide file tree
Changes from 5 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion Project.toml
Original file line number Diff line number Diff line change
Expand Up @@ -122,8 +122,8 @@ RecursiveFactorization = "0.2.23"
Reexport = "1.2.2"
SafeTestsets = "0.1"
SciMLBase = "2.70"
SciMLOperators = "1"
SciMLLogging = "1"
SciMLOperators = "1"
Setfield = "1.1.1"
SparseArrays = "1.10"
Sparspak = "0.3.9"
Expand Down
5 changes: 4 additions & 1 deletion ext/LinearSolveCUDAExt.jl
Original file line number Diff line number Diff line change
@@ -1,13 +1,16 @@
module LinearSolveCUDAExt

using CUDA
using CUDA: CuVector, CuMatrix
using LinearSolve: LinearSolve, is_cusparse, defaultalg, cudss_loaded, DefaultLinearSolver,
DefaultAlgorithmChoice, ALREADY_WARNED_CUDSS, LinearCache,
needs_concrete_A,
error_no_cudss_lu, init_cacheval, OperatorAssumptions,
CudaOffloadFactorization, CudaOffloadLUFactorization, CudaOffloadQRFactorization,
SparspakFactorization, KLUFactorization, UMFPACKFactorization
SparspakFactorization, KLUFactorization, UMFPACKFactorization,
LinearVerbosity
using LinearSolve.LinearAlgebra, LinearSolve.SciMLBase, LinearSolve.ArrayInterface
using LinearAlgebra: LU
using SciMLBase: AbstractSciMLOperator

function LinearSolve.is_cusparse(A::Union{
Expand Down
3 changes: 3 additions & 0 deletions lib/LinearSolveAutotune/Project.toml
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,9 @@ uuid = "67398393-80e8-4254-b7e4-1b9a36a3c5b6"
authors = ["SciML"]
version = "1.8.0"

[sources]
LinearSolve = {path = "../.."}

[deps]
Base64 = "2a0f44e3-6c83-55bd-87e4-b1978d98bd5f"
BenchmarkTools = "6e4b80f9-dd63-53aa-95a3-0cdb28fa8baf"
Expand Down
4 changes: 0 additions & 4 deletions src/factorization.jl
Original file line number Diff line number Diff line change
Expand Up @@ -1210,10 +1210,6 @@ function init_cacheval(::BLISLUFactorization, A, b, u, Pl, Pr,
nothing
end

function init_cacheval(::CudaOffloadLUFactorization, A, b, u, Pl, Pr,
maxiters::Int, abstol, reltol, verbose::LinearVerbosity, assumptions::OperatorAssumptions)
nothing
end

function init_cacheval(::MetalLUFactorization, A, b, u, Pl, Pr,
maxiters::Int, abstol, reltol, verbose::LinearVerbosity, assumptions::OperatorAssumptions)
Expand Down
Loading