Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: safetensors_rust.SafetensorError: device privateuseone:0 is invalid #559

Open
4 of 6 tasks
Bashlator opened this issue Nov 3, 2024 · 5 comments
Open
4 of 6 tasks
Labels
can't reproduce The maintainer could not reproduce this issue. directml DirectML related or specific issue navi

Comments

@Bashlator
Copy link

Checklist

  • The issue exists after disabling all extensions
  • The issue exists on a clean installation of webui
  • The issue is caused by an extension, but I believe it is caused by a bug in the webui
  • The issue exists in the current version of the webui
  • The issue has not been reported before recently
  • The issue has been reported before but has not been fixed yet

What happened?

When trying to run any other model than Stable Diffusion 1.5 (Pony Diffusion V6XL) i get safetensor error

Steps to reproduce the problem

  1. Open webui-user.bat
  2. Wait until webui opens
  3. Under "Stable Diffusion checkpoint" choose Pony Diffusion

What should have happened?

WebUI should load the model

What browsers do you use to access the UI ?

Other

Sysinfo

sysinfo-2024-11-03-14-59.json

Console logs

venv "D:\stable-diffusion-webui-amdgpu\venv\Scripts\Python.exe"
Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug  1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)]
Version: v1.10.1-amd-16-g4730df18
Commit hash: 4730df185b557f1453a0f5f79ffd1fa7b36aae54
D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\timm\models\layers\__init__.py:48: FutureWarning: Importing from timm.models.layers is deprecated, please import via timm.layers
  warnings.warn(f"Importing from {__name__} is deprecated, please import via timm.layers", FutureWarning)
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\pytorch_lightning\utilities\distributed.py:258: LightningDeprecationWarning: `pytorch_lightning.utilities.distributed.rank_zero_only` has been deprecated in v1.8.1 and will be removed in v2.0.0. You can import it from `pytorch_lightning.utilities` instead.
  rank_zero_deprecation(
Launching Web UI with arguments: --use-directml --medvram --lowram
ONNX: version=1.20.0 provider=DmlExecutionProvider, available=['DmlExecutionProvider', 'CPUExecutionProvider']
Calculating sha256 for D:\stable-diffusion-webui-amdgpu\models\Stable-diffusion\ponyDiffusionV6XL_v6StartWithThisOne.safetensors: Running on local URL:  http://127.0.0.1:7860

To create a public link, set `share=True` in `launch()`.
Startup time: 15.8s (prepare environment: 17.3s, initialize shared: 2.1s, other imports: 0.4s, list SD models: 1.9s, load scripts: 1.3s, create ui: 0.5s, gradio launch: 0.4s).
67ab2fd8ec439a89b3fedb15cc65f54336af163c7eb5e4f2acc98f090a29b0b3
Loading weights [67ab2fd8ec] from D:\stable-diffusion-webui-amdgpu\models\Stable-diffusion\ponyDiffusionV6XL_v6StartWithThisOne.safetensors
loading stable diffusion model: SafetensorError
Traceback (most recent call last):
  File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 973, in _bootstrap
    self._bootstrap_inner()
  File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 953, in run
    self._target(*self._args, **self._kwargs)
  File "D:\stable-diffusion-webui-amdgpu\modules\initialize.py", line 149, in load_model
    shared.sd_model  # noqa: B018
  File "D:\stable-diffusion-webui-amdgpu\modules\shared_items.py", line 190, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 693, in get_sd_model
    load_model()
  File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 813, in load_model
    state_dict = get_checkpoint_state_dict(checkpoint_info, timer)
  File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 344, in get_checkpoint_state_dict
    res = read_state_dict(checkpoint_info.filename)
  File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 318, in read_state_dict
    pl_sd = safetensors.torch.load_file(checkpoint_file, device=device)
  File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\safetensors\torch.py", line 308, in load_file
    with safe_open(filename, framework="pt", device=device) as f:
safetensors_rust.SafetensorError: device privateuseone:0 is invalid


Stable diffusion model failed to load
Applying attention optimization: InvokeAI... done.
Loading weights [67ab2fd8ec] from D:\stable-diffusion-webui-amdgpu\models\Stable-diffusion\ponyDiffusionV6XL_v6StartWithThisOne.safetensors
loading stable diffusion model: SafetensorError
Traceback (most recent call last):
  File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 973, in _bootstrap
    self._bootstrap_inner()
  File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\gradio\utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "D:\stable-diffusion-webui-amdgpu\modules\ui.py", line 1740, in <lambda>
    visible=shared.sd_model and shared.sd_model.cond_stage_key == "edit"
  File "D:\stable-diffusion-webui-amdgpu\modules\shared_items.py", line 190, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 693, in get_sd_model
    load_model()
  File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 813, in load_model
    state_dict = get_checkpoint_state_dict(checkpoint_info, timer)
  File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 344, in get_checkpoint_state_dict
    res = read_state_dict(checkpoint_info.filename)
  File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 318, in read_state_dict
    pl_sd = safetensors.torch.load_file(checkpoint_file, device=device)
  File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\safetensors\torch.py", line 308, in load_file
    with safe_open(filename, framework="pt", device=device) as f:
safetensors_rust.SafetensorError: device privateuseone:0 is invalid


Stable diffusion model failed to load
Loading weights [67ab2fd8ec] from D:\stable-diffusion-webui-amdgpu\models\Stable-diffusion\ponyDiffusionV6XL_v6StartWithThisOne.safetensors
changing setting sd_model_checkpoint to ponyDiffusionV6XL_v6StartWithThisOne.safetensors: SafetensorError
Traceback (most recent call last):
  File "D:\stable-diffusion-webui-amdgpu\modules\options.py", line 165, in set
    option.onchange()
  File "D:\stable-diffusion-webui-amdgpu\modules\call_queue.py", line 14, in f
    res = func(*args, **kwargs)
  File "D:\stable-diffusion-webui-amdgpu\modules\initialize_util.py", line 181, in <lambda>
    shared.opts.onchange("sd_model_checkpoint", wrap_queued_call(lambda: sd_models.reload_model_weights()), call=False)
  File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 982, in reload_model_weights
    state_dict = get_checkpoint_state_dict(checkpoint_info, timer)
  File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 344, in get_checkpoint_state_dict
    res = read_state_dict(checkpoint_info.filename)
  File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 318, in read_state_dict
    pl_sd = safetensors.torch.load_file(checkpoint_file, device=device)
  File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\safetensors\torch.py", line 308, in load_file
    with safe_open(filename, framework="pt", device=device) as f:
safetensors_rust.SafetensorError: device privateuseone:0 is invalid

Additional information

I have 16 GB of RAM and RX 5700 with 8 GB VRAM, also using directml

@CS1o
Copy link

CS1o commented Nov 3, 2024

Hey, with DirectML its really hard to get SDXL/Pony models working as it requires a lot more vram than usual.
The best option for you is to use the ZLUDA backend for the Webui.
You can find an install guide here:
https://github.com/CS1o/Stable-Diffusion-Info/wiki/Webui-Installation-Guides
Follow the Automatic1111 with ZLUDA Guide.

@Bashlator
Copy link
Author

Hey, with DirectML its really hard to get SDXL/Pony models working as it requires a lot more vram than usual. The best option for you is to use the ZLUDA backend for the Webui. You can find an install guide here: https://github.com/CS1o/Stable-Diffusion-Info/wiki/Webui-Installation-Guides Follow the Automatic1111 with ZLUDA Guide.

Did everything as in tutorial, now when i try to choose the model entire webui just crashes

venv "D:\stable-diffusion-webui-amdgpu\venv\Scripts\Python.exe"
WARNING: ZLUDA works best with SD.Next. Please consider migrating to SD.Next.
Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)]
Version: v1.10.1-amd-16-g4730df18
Commit hash: 4730df1
ROCm: agents=['gfx1010:xnack-']
ROCm: version=6.1, using agent gfx1010:xnack-
ZLUDA support: experimental
Using ZLUDA in D:\stable-diffusion-webui-amdgpu.zluda
D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\timm\models\layers_init_.py:48: FutureWarning: Importing from timm.models.layers is deprecated, please import via timm.layers
warnings.warn(f"Importing from {name} is deprecated, please import via timm.layers", FutureWarning)
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\pytorch_lightning\utilities\distributed.py:258: LightningDeprecationWarning: pytorch_lightning.utilities.distributed.rank_zero_only has been deprecated in v1.8.1 and will be removed in v2.0.0. You can import it from pytorch_lightning.utilities instead.
rank_zero_deprecation(
Launching Web UI with arguments: --use-zluda --medvram --lowram
Warning: caught exception 'Torch not compiled with CUDA enabled', memory monitor disabled
ONNX: version=1.20.0 provider=CUDAExecutionProvider, available=['TensorrtExecutionProvider', 'CUDAExecutionProvider', 'CPUExecutionProvider']
Loading weights [67ab2fd8ec] from D:\stable-diffusion-webui-amdgpu\models\Stable-diffusion\ponyDiffusionV6XL_v6StartWithThisOne.safetensors
Running on local URL: http://127.0.0.1:7860

To create a public link, set share=True in launch().
Creating model from config: D:\stable-diffusion-webui-amdgpu\repositories\generative-models\configs\inference\sd_xl_base.yaml
Startup time: 14.3s (prepare environment: 20.6s, initialize shared: 1.0s, list SD models: 0.5s, load scripts: 0.7s, create ui: 0.5s, gradio launch: 0.5s).
creating model quickly: OSError
Traceback (most recent call last):
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\utils_http.py", line 406, in hf_raise_for_status
response.raise_for_status()
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\requests\models.py", line 1024, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 401 Client Error: Unauthorized for url: https://huggingface.co/None/resolve/main/config.json

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\transformers\utils\hub.py", line 403, in cached_file
resolved_file = hf_hub_download(
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\utils_validators.py", line 114, in _inner_fn
return fn(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\file_download.py", line 862, in hf_hub_download
return _hf_hub_download_to_cache_dir(
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\file_download.py", line 969, in _hf_hub_download_to_cache_dir
_raise_on_head_call_error(head_call_error, force_download, local_files_only)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\file_download.py", line 1484, in _raise_on_head_call_error
raise head_call_error
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\file_download.py", line 1376, in _get_metadata_or_catch_error
metadata = get_hf_file_metadata(
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\utils_validators.py", line 114, in _inner_fn
return fn(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\file_download.py", line 1296, in get_hf_file_metadata
r = _request_wrapper(
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\file_download.py", line 277, in _request_wrapper
response = _request_wrapper(
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\file_download.py", line 301, in _request_wrapper
hf_raise_for_status(response)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\utils_http.py", line 454, in hf_raise_for_status
raise _format(RepositoryNotFoundError, message, response) from e
huggingface_hub.errors.RepositoryNotFoundError: 401 Client Error. (Request ID: Root=1-6727d028-29f2248650f75c05602abfb0;109edd08-8add-4ce7-8177-24f4cd7bb744)

Repository Not Found for url: https://huggingface.co/None/resolve/main/config.json.
Please make sure you specified the correct repo_id and repo_type.
If you are trying to access a private or gated repo, make sure you are authenticated.
Invalid username or password.

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 973, in _bootstrap
self._bootstrap_inner()
File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 1016, in _bootstrap_inner
self.run()
File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 953, in run
self._target(*self._args, **self._kwargs)
File "D:\stable-diffusion-webui-amdgpu\modules\initialize.py", line 149, in load_model
shared.sd_model # noqa: B018
File "D:\stable-diffusion-webui-amdgpu\modules\shared_items.py", line 190, in sd_model
return modules.sd_models.model_data.get_sd_model()
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 693, in get_sd_model
load_model()
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 831, in load_model
sd_model = instantiate_from_config(sd_config.model, state_dict)
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 775, in instantiate_from_config
return constructor(**params)
File "D:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\models\diffusion.py", line 61, in init
self.conditioner = instantiate_from_config(
File "D:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\util.py", line 175, in instantiate_from_config
return get_obj_from_str(config["target"])(**config.get("params", dict()))
File "D:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\modules\encoders\modules.py", line 88, in init
embedder = instantiate_from_config(embconfig)
File "D:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\util.py", line 175, in instantiate_from_config
return get_obj_from_str(config["target"])(**config.get("params", dict()))
File "D:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\modules\encoders\modules.py", line 361, in init
self.transformer = CLIPTextModel.from_pretrained(version)
File "D:\stable-diffusion-webui-amdgpu\modules\sd_disable_initialization.py", line 68, in CLIPTextModel_from_pretrained
res = self.CLIPTextModel_from_pretrained(None, *model_args, config=pretrained_model_name_or_path, state_dict={}, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\transformers\modeling_utils.py", line 3506, in from_pretrained
resolved_config_file = cached_file(
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\transformers\utils\hub.py", line 426, in cached_file
raise EnvironmentError(
OSError: None is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models'
If this is a private repository, make sure to pass a token having permission to this repo either by logging in with huggingface-cli login or by passing token=<your_token>

Failed to create model quickly; will retry using slow method.
D:\stable-diffusion-webui-amdgpu\modules\safe.py:156: FutureWarning: You are using torch.load with weights_only=False (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for weights_only will be flipped to True. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via torch.serialization.add_safe_globals. We recommend you start setting weights_only=True for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
return unsafe_torch_load(filename, *args, **kwargs)
Loading weights [67ab2fd8ec] from D:\stable-diffusion-webui-amdgpu\models\Stable-diffusion\ponyDiffusionV6XL_v6StartWithThisOne.safetensors
Для продолжения нажмите любую клавишу . . .

@CS1o
Copy link

CS1o commented Nov 3, 2024

Hey, as you only have 16GB of RAM you need to increase the Windows Pagefile.
Instructions here: https://www.tomshardware.com/news/how-to-manage-virtual-memory-pagefile-windows-10,36929.html
Enable it only for the C drive and disable it for any other drive.
Set it to customized: 16000 min and 24000max.
Then restart the PC and try to load the model again.

@Bashlator
Copy link
Author

Hey, as you only have 16GB of RAM you need to increase the Windows Pagefile. Instructions here: https://www.tomshardware.com/news/how-to-manage-virtual-memory-pagefile-windows-10,36929.html Enable it only for the C drive and disable it for any other drive. Set it to customized: 16000 min and 24000max. Then restart the PC and try to load the model again.

I don't even have 24000 left on my SSD. I guess I'll just use another model

@Bashlator
Copy link
Author

Hey, as you only have 16GB of RAM you need to increase the Windows Pagefile. Instructions here: https://www.tomshardware.com/news/how-to-manage-virtual-memory-pagefile-windows-10,36929.html Enable it only for the C drive and disable it for any other drive. Set it to customized: 16000 min and 24000max. Then restart the PC and try to load the model again.

I don't even have 24000 left on my SSD. I guess I'll just use another model

Nevermind, even waifu diffusion isn't loading

venv "D:\stable-diffusion-webui-amdgpu\venv\Scripts\Python.exe"
WARNING: ZLUDA works best with SD.Next. Please consider migrating to SD.Next.
Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)]
Version: v1.10.1-amd-16-g4730df18
Commit hash: 4730df1
ROCm: agents=['gfx1010:xnack-']
ROCm: version=6.1, using agent gfx1010:xnack-
ZLUDA support: experimental
Using ZLUDA in D:\stable-diffusion-webui-amdgpu.zluda
D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\timm\models\layers_init_.py:48: FutureWarning: Importing from timm.models.layers is deprecated, please import via timm.layers
warnings.warn(f"Importing from {name} is deprecated, please import via timm.layers", FutureWarning)
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\pytorch_lightning\utilities\distributed.py:258: LightningDeprecationWarning: pytorch_lightning.utilities.distributed.rank_zero_only has been deprecated in v1.8.1 and will be removed in v2.0.0. You can import it from pytorch_lightning.utilities instead.
rank_zero_deprecation(
Launching Web UI with arguments: --use-zluda --medvram --lowram
Warning: caught exception 'Torch not compiled with CUDA enabled', memory monitor disabled
ONNX: version=1.20.0 provider=CUDAExecutionProvider, available=['TensorrtExecutionProvider', 'CUDAExecutionProvider', 'CPUExecutionProvider']
Loading weights [c76e0962bc] from D:\stable-diffusion-webui-amdgpu\models\Stable-diffusion\wd-1-4-anime_e2.ckpt
D:\stable-diffusion-webui-amdgpu\modules\safe.py:156: FutureWarning: You are using torch.load with weights_only=False (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for weights_only will be flipped to True. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via torch.serialization.add_safe_globals. We recommend you start setting weights_only=True for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
return unsafe_torch_load(filename, *args, **kwargs)
Running on local URL: http://127.0.0.1:7860

To create a public link, set share=True in launch().
Startup time: 39.7s (prepare environment: 63.8s, initialize shared: 3.0s, list SD models: 0.2s, load scripts: 1.4s, create ui: 0.6s, gradio launch: 0.5s).
loading stable diffusion model: RuntimeError
Traceback (most recent call last):
File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 973, in _bootstrap
self._bootstrap_inner()
File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 1016, in _bootstrap_inner
self.run()
File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 953, in run
self._target(*self._args, **self._kwargs)
File "D:\stable-diffusion-webui-amdgpu\modules\initialize.py", line 149, in load_model
shared.sd_model # noqa: B018
File "D:\stable-diffusion-webui-amdgpu\modules\shared_items.py", line 190, in sd_model
return modules.sd_models.model_data.get_sd_model()
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 693, in get_sd_model
load_model()
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 815, in load_model
checkpoint_config = sd_models_config.find_checkpoint_config(state_dict, checkpoint_info)
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models_config.py", line 125, in find_checkpoint_config
return guess_model_config_from_state_dict(state_dict, info.filename)
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models_config.py", line 98, in guess_model_config_from_state_dict
elif is_using_v_parameterization_for_sd2(sd):
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models_config.py", line 67, in is_using_v_parameterization_for_sd2
out = (unet(x_test, torch.asarray([999], device=device), context=test_cond) - x_test).mean().cpu().item()
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\modules\sd_unet.py", line 91, in UNetModel_forward
return original_forward(self, x, timesteps, context, *args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\repositories\stable-diffusion-stability-ai\ldm\modules\diffusionmodules\openaimodel.py", line 797, in forward
h = module(h, emb, context)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\repositories\stable-diffusion-stability-ai\ldm\modules\diffusionmodules\openaimodel.py", line 86, in forward
x = layer(x)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\extensions-builtin\Lora\networks.py", line 599, in network_Conv2d_forward
return originals.Conv2d_forward(self, input)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\conv.py", line 458, in forward
return self._conv_forward(input, self.weight, self.bias)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\conv.py", line 454, in _conv_forward
return F.conv2d(input, weight, bias, self.stride,
RuntimeError: Input type (float) and bias type (struct c10::Half) should be the same

Stable diffusion model failed to load
Applying attention optimization: InvokeAI... done.
Loading weights [c76e0962bc] from D:\stable-diffusion-webui-amdgpu\models\Stable-diffusion\wd-1-4-anime_e2.ckpt
D:\stable-diffusion-webui-amdgpu\modules\safe.py:156: FutureWarning: You are using torch.load with weights_only=False (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for weights_only will be flipped to True. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via torch.serialization.add_safe_globals. We recommend you start setting weights_only=True for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
return unsafe_torch_load(filename, *args, **kwargs)
loading stable diffusion model: RuntimeError
Traceback (most recent call last):
File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 973, in _bootstrap
self._bootstrap_inner()
File "C:\Users\vasil\AppData\Local\Programs\Python\Python310\lib\threading.py", line 1016, in _bootstrap_inner
self.run()
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\anyio_backends_asyncio.py", line 807, in run
result = context.run(func, *args)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\gradio\utils.py", line 707, in wrapper
response = f(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\modules\ui.py", line 1740, in
visible=shared.sd_model and shared.sd_model.cond_stage_key == "edit"
File "D:\stable-diffusion-webui-amdgpu\modules\shared_items.py", line 190, in sd_model
return modules.sd_models.model_data.get_sd_model()
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 693, in get_sd_model
load_model()
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 815, in load_model
checkpoint_config = sd_models_config.find_checkpoint_config(state_dict, checkpoint_info)
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models_config.py", line 125, in find_checkpoint_config
return guess_model_config_from_state_dict(state_dict, info.filename)
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models_config.py", line 98, in guess_model_config_from_state_dict
elif is_using_v_parameterization_for_sd2(sd):
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models_config.py", line 67, in is_using_v_parameterization_for_sd2
out = (unet(x_test, torch.asarray([999], device=device), context=test_cond) - x_test).mean().cpu().item()
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\modules\sd_unet.py", line 91, in UNetModel_forward
return original_forward(self, x, timesteps, context, *args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\repositories\stable-diffusion-stability-ai\ldm\modules\diffusionmodules\openaimodel.py", line 797, in forward
h = module(h, emb, context)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\repositories\stable-diffusion-stability-ai\ldm\modules\diffusionmodules\openaimodel.py", line 86, in forward
x = layer(x)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\extensions-builtin\Lora\networks.py", line 599, in network_Conv2d_forward
return originals.Conv2d_forward(self, input)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\conv.py", line 458, in forward
return self._conv_forward(input, self.weight, self.bias)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\conv.py", line 454, in _conv_forward
return F.conv2d(input, weight, bias, self.stride,
RuntimeError: Input type (float) and bias type (struct c10::Half) should be the same

Stable diffusion model failed to load
Loading weights [c76e0962bc] from D:\stable-diffusion-webui-amdgpu\models\Stable-diffusion\wd-1-4-anime_e2.ckpt
changing setting sd_model_checkpoint to wd-1-4-anime_e2.ckpt [c76e0962bc]: RuntimeError
Traceback (most recent call last):
File "D:\stable-diffusion-webui-amdgpu\modules\options.py", line 165, in set
option.onchange()
File "D:\stable-diffusion-webui-amdgpu\modules\call_queue.py", line 14, in f
res = func(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\modules\initialize_util.py", line 181, in
shared.opts.onchange("sd_model_checkpoint", wrap_queued_call(lambda: sd_models.reload_model_weights()), call=False)
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models.py", line 984, in reload_model_weights
checkpoint_config = sd_models_config.find_checkpoint_config(state_dict, checkpoint_info)
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models_config.py", line 125, in find_checkpoint_config
return guess_model_config_from_state_dict(state_dict, info.filename)
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models_config.py", line 98, in guess_model_config_from_state_dict
elif is_using_v_parameterization_for_sd2(sd):
File "D:\stable-diffusion-webui-amdgpu\modules\sd_models_config.py", line 67, in is_using_v_parameterization_for_sd2
out = (unet(x_test, torch.asarray([999], device=device), context=test_cond) - x_test).mean().cpu().item()
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\modules\sd_unet.py", line 91, in UNetModel_forward
return original_forward(self, x, timesteps, context, *args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\repositories\stable-diffusion-stability-ai\ldm\modules\diffusionmodules\openaimodel.py", line 797, in forward
h = module(h, emb, context)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\repositories\stable-diffusion-stability-ai\ldm\modules\diffusionmodules\openaimodel.py", line 86, in forward
x = layer(x)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
File "D:\stable-diffusion-webui-amdgpu\extensions-builtin\Lora\networks.py", line 599, in network_Conv2d_forward
return originals.Conv2d_forward(self, input)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\conv.py", line 458, in forward
return self._conv_forward(input, self.weight, self.bias)
File "D:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\conv.py", line 454, in _conv_forward
return F.conv2d(input, weight, bias, self.stride,
RuntimeError: Input type (float) and bias type (struct c10::Half) should be the same

@lshqqytiger lshqqytiger added directml DirectML related or specific issue can't reproduce The maintainer could not reproduce this issue. navi labels Nov 4, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
can't reproduce The maintainer could not reproduce this issue. directml DirectML related or specific issue navi
Projects
None yet
Development

No branches or pull requests

3 participants