InvokeAI: [bug]: Fresh install always uses cuda for a ROCm compatible AMD GPU
Is there an existing issue for this?
- I have searched the existing issues
OS
Linux
GPU
amd
VRAM
16GB
What version did you experience this issue on?
v3.0.2rc1
What happened?
I used the install script from the latest release, and selected AMD GPU (with ROCm). The script installs perfectly fine, and then I go to run InvokeAI in the graphical web client, and I get this output:
[2023-08-09 17:25:18,419]::[uvicorn.error]::INFO --> Started server process [16866]
[2023-08-09 17:25:18,419]::[uvicorn.error]::INFO --> Waiting for application startup.
[2023-08-09 17:25:18,420]::[InvokeAI]::INFO --> InvokeAI version 3.0.1
[2023-08-09 17:25:18,420]::[InvokeAI]::INFO --> Root directory = /<myrootpath>/AI/invokeAI
[2023-08-09 17:25:18,421]::[InvokeAI]::INFO --> GPU device = cuda AMD Radeon RX 6800 XT
As you can see, it opens up with cuda AMD Radeon RX 6800 XT. This card works just fine with A111 and ROCm. I’ve also edited the invokeai.yaml file, as I saw that xformers was enabled (isn’t available for AMD cards). Here’s my current config:
InvokeAI:
Web Server:
host: 127.0.0.1
port: 9090
allow_origins: []
allow_credentials: true
allow_methods:
- '*'
allow_headers:
- '*'
Features:
esrgan: true
internet_available: true
log_tokenization: false
patchmatch: true
Memory/Performance:
always_use_cpu: false
free_gpu_mem: true
max_cache_size: 10.0
max_vram_cache_size: 2.75
precision: float32
sequential_guidance: false
xformers_enabled: false
tiled_decode: false
Paths:
autoimport_dir: autoimport
lora_dir: null
embedding_dir: null
controlnet_dir: null
conf_path: configs/models.yaml
models_dir: models
legacy_conf_dir: configs/stable-diffusion
db_dir: databases
outdir: /<myrootpath>/AI/invokeAI/outputs
use_memory_db: false
Logging:
log_handlers:
- console
log_format: color
log_level: info
Of course, cuda doesn’t work with my card and I get all-black output images.
Screenshots
No response
Additional context
This also breaks on the manual install and runs with cuda.
Contact Details
No response
About this issue
- Original URL
- State: open
- Created a year ago
- Reactions: 1
- Comments: 17 (3 by maintainers)
I’m having a similar problem using a 7900 XTX even after I installed Invoke with
pip install InvokeAI --use-pep517 --extra-index-url https://download.pytorch.org/whl/nightly/rocm5.6Invoke still fully installs only the Nvidia stack, then launches in cpu only mode
export HSA_OVERRIDE_GFX_VERSION=10.3.0Before running
./invoke.shappears to have fixed the issue for me on my 6750 xt.7 Series AMD cards may need this:
export HSA_OVERRIDE_GFX_VERSION=11.0.0Also of note, tweaking the version numbers of the above suggested command runs without errors:
pip install "torch==2.1.2+rocm5.6" "torchvision==0.16.2+rocm5.6" "fsspec==2023.10.0.0" "requests~=2.28.2" --force-reinstall --extra-index-url https://download.pytorch.org/whl/rocm5.6I don’t know enough to be able to say if those versions are a good idea, but they don’t error for me on invoke 3.6.0rc6.disclaimer: I don’t have any idea what I’m doing
I have tried to manually modify the create_install.sh and associated files to remove CUDA. I have tried to force ROCM install. I still haven’t figured things out. I wish they had the old requirements.txt file. At this point InvokeAI is useless for AMD GPUs.
I tried to update installer.py with the following: