-
Notifications
You must be signed in to change notification settings - Fork 2.8k
Description
Is there an existing issue for this problem?
- I have searched the existing issues
Install method
Manual
Operating system
Linux
GPU vendor
AMD (ROCm)
GPU model
RX 580
GPU VRAM
8GB
Version number
6.9.0
Browser
Firefox 146.0
System Information
{
"version": "6.9.0",
"dependencies": {
"absl-py" : "2.3.1",
"accelerate" : "1.8.1",
"annotated-types" : "0.7.0",
"anyio" : "4.9.0",
"attrs" : "25.3.0",
"bidict" : "0.23.1",
"bitsandbytes" : "0.46.1",
"blake3" : "1.0.5",
"certifi" : "2025.6.15",
"cffi" : "1.17.1",
"charset-normalizer" : "3.4.2",
"click" : "8.2.1",
"coloredlogs" : "15.0.1",
"compel" : "2.1.1",
"contourpy" : "1.3.2",
"CUDA" : "N/A",
"cycler" : "0.12.1",
"Deprecated" : "1.2.18",
"diffusers" : "0.33.0",
"dnspython" : "2.7.0",
"dynamicprompts" : "0.31.0",
"einops" : "0.8.1",
"fastapi" : "0.118.3",
"fastapi-events" : "0.12.2",
"filelock" : "3.18.0",
"flatbuffers" : "25.2.10",
"fonttools" : "4.58.5",
"fsspec" : "2025.5.1",
"gguf" : "0.17.1",
"h11" : "0.16.0",
"hf-xet" : "1.1.5",
"httptools" : "0.6.4",
"huggingface-hub" : "0.34.4",
"humanfriendly" : "10.0",
"idna" : "3.10",
"importlib_metadata" : "8.7.0",
"InvokeAI" : "6.9.0",
"jax" : "0.6.2",
"jaxlib" : "0.6.2",
"Jinja2" : "3.1.6",
"kiwisolver" : "1.4.8",
"MarkupSafe" : "3.0.1",
"matplotlib" : "3.10.3",
"mediapipe" : "0.10.14",
"ml_dtypes" : "0.5.1",
"mpmath" : "1.3.0",
"networkx" : "3.5",
"numpy" : "1.26.4",
"onnx" : "1.16.1",
"onnxruntime" : "1.19.2",
"opencv-contrib-python": "4.11.0.86",
"opt_einsum" : "3.4.0",
"packaging" : "25.0",
"picklescan" : "0.0.26",
"pillow" : "11.3.0",
"prompt_toolkit" : "3.0.51",
"protobuf" : "4.25.8",
"psutil" : "7.0.0",
"pycparser" : "2.22",
"pydantic" : "2.11.7",
"pydantic-settings" : "2.10.1",
"pydantic_core" : "2.33.2",
"pyparsing" : "3.2.3",
"PyPatchMatch" : "1.0.2",
"python-dateutil" : "2.9.0.post0",
"python-dotenv" : "1.1.1",
"python-engineio" : "4.12.2",
"python-multipart" : "0.0.20",
"python-socketio" : "5.13.0",
"pytorch-triton-rocm" : "3.3.1",
"PyWavelets" : "1.8.0",
"PyYAML" : "6.0.2",
"regex" : "2024.11.6",
"requests" : "2.32.4",
"safetensors" : "0.5.3",
"scipy" : "1.16.0",
"semver" : "3.0.4",
"sentencepiece" : "0.2.0",
"setuptools" : "80.9.0",
"simple-websocket" : "1.1.0",
"six" : "1.17.0",
"sniffio" : "1.3.1",
"sounddevice" : "0.5.2",
"spandrel" : "0.4.1",
"starlette" : "0.46.2",
"sympy" : "1.14.0",
"tokenizers" : "0.22.0",
"torch" : "2.7.1+rocm6.3",
"torchsde" : "0.2.6",
"torchvision" : "0.22.1+rocm6.3",
"tqdm" : "4.67.1",
"trampoline" : "0.1.2",
"transformers" : "4.56.0",
"typing-inspection" : "0.4.1",
"typing_extensions" : "4.14.0",
"urllib3" : "2.5.0",
"uvicorn" : "0.35.0",
"uvloop" : "0.21.0",
"watchfiles" : "1.1.0",
"wcwidth" : "0.2.13",
"websockets" : "15.0.1",
"wrapt" : "1.17.2",
"wsproto" : "1.2.0",
"zipp" : "3.23.0"
},
"config": {
"schema_version": "4.0.2",
"legacy_models_yaml_path": null,
"host": "0.0.0.0",
"port": 9090,
"allow_origins": [],
"allow_credentials": true,
"allow_methods": [""],
"allow_headers": [""],
"ssl_certfile": null,
"ssl_keyfile": null,
"log_tokenization": false,
"patchmatch": true,
"models_dir": "models",
"convert_cache_dir": "models/.convert_cache",
"download_cache_dir": "models/.download_cache",
"legacy_conf_dir": "configs",
"db_dir": "databases",
"outputs_dir": "outputs",
"custom_nodes_dir": "nodes",
"style_presets_dir": "style_presets",
"workflow_thumbnails_dir": "workflow_thumbnails",
"log_handlers": ["console"],
"log_format": "color",
"log_level": "info",
"log_sql": false,
"log_level_network": "warning",
"use_memory_db": false,
"dev_reload": false,
"profile_graphs": false,
"profile_prefix": null,
"profiles_dir": "profiles",
"max_cache_ram_gb": null,
"max_cache_vram_gb": null,
"log_memory_usage": false,
"device_working_mem_gb": 3,
"enable_partial_loading": false,
"keep_ram_copy_of_weights": true,
"ram": null,
"vram": null,
"lazy_offload": true,
"pytorch_cuda_alloc_conf": null,
"device": "auto",
"precision": "auto",
"sequential_guidance": false,
"attention_type": "auto",
"attention_slice_size": "auto",
"force_tiled_decode": false,
"pil_compress_level": 1,
"max_queue_size": 10000,
"clear_queue_on_startup": false,
"allow_nodes": null,
"deny_nodes": null,
"node_cache_size": 512,
"hashing_algorithm": "blake3_single",
"remote_api_tokens": null,
"scan_models_on_startup": false,
"unsafe_disable_picklescan": false,
"allow_unknown_models": true
},
"set_config_fields": ["host", "legacy_models_yaml_path", "port"]
}
What happened
I'm trying to get ROCm to work in the container, but I get this message in the docker log : The installed version of bitsandbytes was compiled without GPU support
I pulled the ghcr.io/invoke-ai/invokeai:main-rocm image.
What you expected to happen
Being able to use GPU.
How to reproduce the problem
No response
Additional context
No response
Discord username
No response