Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: Mac更新1.6.0后无法使用TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead. #12907

Open
1 task done
owspace opened this issue Aug 31, 2023 · 61 comments
Labels
bug-report Report of a bug, yet to be confirmed platform:mac Issues that apply to Apple OS X, M1, M2, etc

Comments

@owspace
Copy link

owspace commented Aug 31, 2023

Is there an existing issue for this?

  • I have searched the existing issues and checked the recent builds/commits

What happened?

启动webui后无法生成图像

Steps to reproduce the problem

1.启动webui
2. 输入提示词
3.开始生成图片

What should have happened?

正常生成图片

Sysinfo

{
"date": "Thu Aug 31 22:36:37 2023",
"timestamp": "22:36:52",
"uptime": "Thu Aug 31 22:18:16 2023",
"version": {
"app": "stable-diffusion-webui",
"updated": "2023-08-31",
"hash": "5ef669de",
"url": "/~https://github.com/AUTOMATIC1111/stable-diffusion-webui/tree/master"
},
"torch": "2.0.1 autocast half",
"gpu": {},
"state": {
"started": "Thu Aug 31 22:36:52 2023",
"step": "0 / 0",
"jobs": "0 / 0",
"flags": "",
"job": "",
"text-info": ""
},
"memory": {
"ram": {
"free": 29.98,
"used": 2.02,
"total": 32
}
},
"optimizations": [
"none"
],
"libs": {
"xformers": "",
"diffusers": "",
"transformers": "4.30.2"
},
"repos": {
"Stable Diffusion": "[cf1d67a] 2023-03-25",
"Stable Diffusion XL": "[45c443b] 2023-07-26",
"CodeFormer": "[c5b4593] 2022-09-09",
"BLIP": "[48211a1] 2022-06-07",
"k_diffusion": "[ab527a9] 2023-08-12"
},
"device": {
"active": "mps",
"dtype": "torch.float16",
"vae": "torch.float32",
"unet": "torch.float16"
},

}

What browsers do you use to access the UI ?

Google Chrome

Console logs

Loading weights [0b9e46a0b0] from /Users/alyears/stable-diffusion-webui/models/Stable-diffusion/人物/墨幽人造人_v1040.safetensors
Traceback (most recent call last):
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/routes.py", line 488, in run_predict
    output = await app.get_blocks().process_api(
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/blocks.py", line 1431, in process_api
    result = await self.call_function(
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/blocks.py", line 1103, in call_function
    prediction = await anyio.to_thread.run_sync(
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/to_thread.py", line 33, in run_sync
    return await get_asynclib().run_sync_in_worker_thread(
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
    return await future
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/alyears/stable-diffusion-webui/modules/ui_extra_networks.py", line 392, in pages_html
    return refresh()
  File "/Users/alyears/stable-diffusion-webui/modules/ui_extra_networks.py", line 398, in refresh
    pg.refresh()
  File "/Users/alyears/stable-diffusion-webui/modules/ui_extra_networks_textual_inversion.py", line 13, in refresh
    sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
  File "/Users/alyears/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 255, in load_textual_inversion_embeddings
    self.expected_shape = self.get_expected_shape()
  File "/Users/alyears/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 154, in get_expected_shape
    vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
AttributeError: 'NoneType' object has no attribute 'cond_stage_model'
Creating model from config: /Users/alyears/stable-diffusion-webui/configs/v1-inference.yaml
loading stable diffusion model: TypeError
Traceback (most recent call last):
  File "/opt/homebrew/Cellar/python@3.10/3.10.12_1/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 973, in _bootstrap
    self._bootstrap_inner()
  File "/opt/homebrew/Cellar/python@3.10/3.10.12_1/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/alyears/stable-diffusion-webui/modules/ui.py", line 1298, in <lambda>
    update_image_cfg_scale_visibility = lambda: gr.update(visible=shared.sd_model and shared.sd_model.cond_stage_key == "edit")
  File "/Users/alyears/stable-diffusion-webui/modules/shared_items.py", line 110, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Users/alyears/stable-diffusion-webui/modules/sd_models.py", line 499, in get_sd_model
    load_model()
  File "/Users/alyears/stable-diffusion-webui/modules/sd_models.py", line 626, in load_model
    load_model_weights(sd_model, checkpoint_info, state_dict, timer)
  File "/Users/alyears/stable-diffusion-webui/modules/sd_models.py", line 353, in load_model_weights
    model.load_state_dict(state_dict, strict=False)
  File "/Users/alyears/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in <lambda>
    module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs))
  File "/Users/alyears/stable-diffusion-webui/modules/sd_disable_initialization.py", line 221, in load_state_dict
    original(module, state_dict, strict=strict)
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2027, in load_state_dict
    load(self, state_dict)
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  [Previous line repeated 3 more times]
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2009, in load
    module._load_from_state_dict(
  File "/Users/alyears/stable-diffusion-webui/modules/sd_disable_initialization.py", line 226, in <lambda>
    conv2d_load_from_state_dict = self.replace(torch.nn.Conv2d, '_load_from_state_dict', lambda *args, **kwargs: load_from_state_dict(conv2d_load_from_state_dict, *args, **kwargs))
  File "/Users/alyears/stable-diffusion-webui/modules/sd_disable_initialization.py", line 191, in load_from_state_dict
    module._parameters[name] = torch.nn.parameter.Parameter(torch.zeros_like(param, device=device, dtype=dtype), requires_grad=param.requires_grad)
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1780, in zeros_like
    return aten.empty_like.default(
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_ops.py", line 287, in __call__
    return self._op(*args, **kwargs or {})
  File "/Users/alyears/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_refs/__init__.py", line 4254, in empty_like
    return torch.empty_strided(
TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.


Stable diffusion model failed to load

Additional information

No response

@owspace owspace added the bug-report Report of a bug, yet to be confirmed label Aug 31, 2023
@w-e-w w-e-w added the platform:mac Issues that apply to Apple OS X, M1, M2, etc label Aug 31, 2023
@w-e-w
Copy link
Collaborator

w-e-w commented Aug 31, 2023

possibly related #12526
but maybe not the actual cause as this one explicitly use float32 for MPS

@WatashiLoveliver
Copy link

Same here :(

@nicklansley
Copy link

Yes, same error (Macbook Pro, M2 Max, 96GB memory):
"TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead."

It was working up until this commit.

Console log

################################################################
Install script for stable-diffusion + Web UI
Tested on Debian 11 (Bullseye)
################################################################

################################################################
Running on nick user
################################################################

################################################################
Repo already cloned, using it as install directory
################################################################

################################################################
python venv already activate or run without venv: /Users/nick/PycharmProjects/stable-diffusion-webui/venv
################################################################

################################################################
Launching launch.py...
################################################################
Python 3.10.13 (main, Aug 24 2023, 22:36:46) [Clang 14.0.3 (clang-1403.0.22.14.1)]
Version: v1.6.0
Commit hash: 5ef669d
Fetching updates for Stable Diffusion XL...
Checking out commit for Stable Diffusion XL with hash: 45c443b316737a4ab6e40413d7794a7f5657c19f...
Previous HEAD position was 5c10dee Merge branch 'main' of /~https://github.com/Stability-AI/generative-models into main
HEAD is now at 45c443b Fix license-files setting for project (#71)
Fetching updates for K-diffusion...
Checking out commit for K-diffusion with hash: ab527a9a6d347f364e3d185ba6d714e22d80cb3c...
Previous HEAD position was 51c9778 Add PyTorch 1.12.1 MPS workaround for DPM fast
HEAD is now at ab527a9 Release 0.0.16
Installing requirements
Launching Web UI with arguments: --skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
2023-09-01 08:41:25.298 Python[5358:226581] apply_selection_policy_once: avoid use of removable GPUs (via org.python.python:GPUSelectionPolicy->avoidRemovable)
Warning: caught exception 'Torch not compiled with CUDA enabled', memory monitor disabled
Loading weights [7ed60a2f58] from /Users/nick/PycharmProjects/stable-diffusion-webui/models/Stable-diffusion/juggernaut_aftermath.safetensors
Creating model from config: /Users/nick/PycharmProjects/stable-diffusion-webui/models/Stable-diffusion/juggernaut_aftermath.yaml
Running on local URL: http://127.0.0.1:7860

To create a public link, set share=True in launch().
Startup time: 26.1s (prepare environment: 14.8s, import torch: 3.0s, import gradio: 1.5s, setup paths: 1.0s, initialize shared: 1.0s, other imports: 3.8s, load scripts: 0.3s, initialize extra networks: 0.2s, create ui: 0.2s, gradio launch: 0.3s).
loading stable diffusion model: TypeError
Traceback (most recent call last):
File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 973, in _bootstrap
self._bootstrap_inner()
File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
self.run()
File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 953, in run
self._target(*self._args, **self._kwargs)
File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/initialize.py", line 147, in load_model
shared.sd_model # noqa: B018
File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/shared_items.py", line 110, in sd_model
return modules.sd_models.model_data.get_sd_model()
File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 499, in get_sd_model
load_model()
File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 626, in load_model
load_model_weights(sd_model, checkpoint_info, state_dict, timer)
File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 353, in load_model_weights
model.load_state_dict(state_dict, strict=False)
File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in
module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs))
File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 221, in load_state_dict
original(module, state_dict, strict=strict)
File "/opt/homebrew/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2027, in load_state_dict
load(self, state_dict)
File "/opt/homebrew/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
load(child, child_state_dict, child_prefix)
File "/opt/homebrew/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
load(child, child_state_dict, child_prefix)
File "/opt/homebrew/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
load(child, child_state_dict, child_prefix)
[Previous line repeated 3 more times]
File "/opt/homebrew/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2009, in load
module._load_from_state_dict(
File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 226, in
conv2d_load_from_state_dict = self.replace(torch.nn.Conv2d, '_load_from_state_dict', lambda *args, **kwargs: load_from_state_dict(conv2d_load_from_state_dict, *args, **kwargs))
File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 191, in load_from_state_dict
module._parameters[name] = torch.nn.parameter.Parameter(torch.zeros_like(param, device=device, dtype=dtype), requires_grad=param.requires_grad)
File "/opt/homebrew/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1780, in zeros_like
return aten.empty_like.default(
File "/opt/homebrew/lib/python3.10/site-packages/torch/_ops.py", line 287, in call
return self._op(*args, **kwargs or {})
File "/opt/homebrew/lib/python3.10/site-packages/torch/_refs/init.py", line 4254, in empty_like
return torch.empty_strided(
TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.

@ngtongsheng
Copy link

ngtongsheng commented Sep 1, 2023

I faced the same issue, the following works for me
- Delete webui-maco-env.sh
- Remove export PYTORCH_MPS_HIGH_WATERMARK_RATIO="0.0" in webui-user.sh if exist
- Add this env export COMMANDLINE_ARGS="--skip-torch-cuda-test --no-half --use-cpu all"

Updated this thread, it was a temporary workaround to force using cpu, you should try @ericwagner101 resolution instead
#12907 (comment)

@akx
Copy link
Collaborator

akx commented Sep 1, 2023

FWIW, on v1.6.0-78-gd39440bf --skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate works fine on my M2 Mac and is using the GPU with both SD and SDXL models.

@nicklansley
Copy link

Thanks @akx but my console output shows those startup settings are executed (14 lines below "Launching launch.py..." in my comment above).

@nicklansley
Copy link

@ngtongsheng Yes that works BUT now it's generating using the actual CPUs and not Apple's Metal GPU!

In this screenshot of Activity Monitor you can see 'Python' with 338% CPU and 0% GPU, and the CPU and GPU graphs confirming this, during the creation of one 512x512 image that took 3.5 seconds per iteration (not bad for a CPU but overall, very slow!).

Screenshot 2023-09-01 at 14 52 54

@akx
Copy link
Collaborator

akx commented Sep 1, 2023

@nicklansley Well, for one you're evidently (/opt/homebrew/lib/python3.10/site-packages/torch/nn/modules/module.py is not a virtualenv path) not (fully) using a virtualenv to run the project (which is really the only correct way to run it, unless you're in a container or some other isolation, but you're not), so chances are your global Python environment (/opt/homebrew/lib/python3.10/site-packages) has wrong versions of some packages.

I would honestly recommend uninstalling all extra packages from that global Python environment, and then ensure you're using a virtualenv going forward.

@nicklansley
Copy link

nicklansley commented Sep 1, 2023

Thanks @akx but I promise you I am using a virtualenv!
From the console log above:
python venv already activate or run without venv: /Users/nick/PycharmProjects/stable-diffusion-webui/venv

HOWEVER... it may be it is mixing up virtualenv and the global site-packages?

OK a way forward - thank you.

@akx
Copy link
Collaborator

akx commented Sep 1, 2023

You are using a virtualenv, but you're clearly using a torch from outside that virtualenv. Without a venv activated, run python3.10 -m pip list and then python3.10 -m pip uninstall -y ... the packages that shouldn't be installed globally (practically everything but pip, wheel, setuptools, pretty much).

@nicklansley
Copy link

nicklansley commented Sep 1, 2023

I wiped everything but the model directory, performed a git clone then let webui.sh create the virtual env and install the packages.

Still the same error but, because i know all the models worked before this commit, I will rollback a couple of weeks and step forward again.

My new run of webui.sh is below, with references to /venv/ packages clearly showing in the execution except for access to /opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py
..which is where homebrew installed the 'master' copy of python@3.10

Perhaps core packages such as threading.py are not copied into the virtual environment when a full path including major.minor.patch (3.10.13) version of Python is used?

Here's the run:

/usr/bin/env bash /Users/nick/PycharmProjects/stable-diffusion-webui/webui.sh

################################################################
Install script for stable-diffusion + Web UI
Tested on Debian 11 (Bullseye)
################################################################

################################################################
Running on nick user
################################################################

################################################################
Repo already cloned, using it as install directory
################################################################

################################################################
Create and activate python venv
################################################################

################################################################
Launching launch.py...
################################################################
Python 3.10.13 (main, Aug 24 2023, 22:36:46) [Clang 14.0.3 (clang-1403.0.22.14.1)]
Version: v1.6.0
Commit hash: 5ef669de080814067961f28357256e8fe27544f4
Installing torch and torchvision
Collecting torch==2.0.1
  Using cached torch-2.0.1-cp310-none-macosx_11_0_arm64.whl (55.8 MB)
Collecting torchvision==0.15.2
  Using cached torchvision-0.15.2-cp310-cp310-macosx_11_0_arm64.whl (1.4 MB)
Collecting filelock (from torch==2.0.1)
  Obtaining dependency information for filelock from https://files.pythonhosted.org/packages/52/90/45223db4e1df30ff14e8aebf9a1bf0222da2e7b49e53692c968f36817812/filelock-3.12.3-py3-none-any.whl.metadata
  Downloading filelock-3.12.3-py3-none-any.whl.metadata (2.7 kB)
Collecting typing-extensions (from torch==2.0.1)
  Obtaining dependency information for typing-extensions from https://files.pythonhosted.org/packages/ec/6b/63cc3df74987c36fe26157ee12e09e8f9db4de771e0f3404263117e75b95/typing_extensions-4.7.1-py3-none-any.whl.metadata
  Downloading typing_extensions-4.7.1-py3-none-any.whl.metadata (3.1 kB)
Collecting sympy (from torch==2.0.1)
  Using cached sympy-1.12-py3-none-any.whl (5.7 MB)
Collecting networkx (from torch==2.0.1)
  Using cached networkx-3.1-py3-none-any.whl (2.1 MB)
Collecting jinja2 (from torch==2.0.1)
  Using cached Jinja2-3.1.2-py3-none-any.whl (133 kB)
Collecting numpy (from torchvision==0.15.2)
  Obtaining dependency information for numpy from https://files.pythonhosted.org/packages/c3/ea/1d95b399078ecaa7b5d791e1fdbb3aee272077d9fd5fb499593c87dec5ea/numpy-1.25.2-cp310-cp310-macosx_11_0_arm64.whl.metadata
  Downloading numpy-1.25.2-cp310-cp310-macosx_11_0_arm64.whl.metadata (5.6 kB)
Collecting requests (from torchvision==0.15.2)
  Obtaining dependency information for requests from https://files.pythonhosted.org/packages/70/8e/0e2d847013cb52cd35b38c009bb167a1a26b2ce6cd6965bf26b47bc0bf44/requests-2.31.0-py3-none-any.whl.metadata
  Downloading requests-2.31.0-py3-none-any.whl.metadata (4.6 kB)
Collecting pillow!=8.3.*,>=5.3.0 (from torchvision==0.15.2)
  Obtaining dependency information for pillow!=8.3.*,>=5.3.0 from https://files.pythonhosted.org/packages/ef/53/024e161112beb11008d6c7529c954e2ec641ae17b99e03fe9a539e114ae6/Pillow-10.0.0-cp310-cp310-macosx_11_0_arm64.whl.metadata
  Downloading Pillow-10.0.0-cp310-cp310-macosx_11_0_arm64.whl.metadata (9.5 kB)
Collecting MarkupSafe>=2.0 (from jinja2->torch==2.0.1)
  Obtaining dependency information for MarkupSafe>=2.0 from https://files.pythonhosted.org/packages/20/1d/713d443799d935f4d26a4f1510c9e61b1d288592fb869845e5cc92a1e055/MarkupSafe-2.1.3-cp310-cp310-macosx_10_9_universal2.whl.metadata
  Downloading MarkupSafe-2.1.3-cp310-cp310-macosx_10_9_universal2.whl.metadata (3.0 kB)
Collecting charset-normalizer<4,>=2 (from requests->torchvision==0.15.2)
  Obtaining dependency information for charset-normalizer<4,>=2 from https://files.pythonhosted.org/packages/ec/a7/96835706283d63fefbbbb4f119d52f195af00fc747e67cc54397c56312c8/charset_normalizer-3.2.0-cp310-cp310-macosx_11_0_arm64.whl.metadata
  Downloading charset_normalizer-3.2.0-cp310-cp310-macosx_11_0_arm64.whl.metadata (31 kB)
Collecting idna<4,>=2.5 (from requests->torchvision==0.15.2)
  Using cached idna-3.4-py3-none-any.whl (61 kB)
Collecting urllib3<3,>=1.21.1 (from requests->torchvision==0.15.2)
  Obtaining dependency information for urllib3<3,>=1.21.1 from https://files.pythonhosted.org/packages/9b/81/62fd61001fa4b9d0df6e31d47ff49cfa9de4af03adecf339c7bc30656b37/urllib3-2.0.4-py3-none-any.whl.metadata
  Downloading urllib3-2.0.4-py3-none-any.whl.metadata (6.6 kB)
Collecting certifi>=2017.4.17 (from requests->torchvision==0.15.2)
  Obtaining dependency information for certifi>=2017.4.17 from https://files.pythonhosted.org/packages/4c/dd/2234eab22353ffc7d94e8d13177aaa050113286e93e7b40eae01fbf7c3d9/certifi-2023.7.22-py3-none-any.whl.metadata
  Downloading certifi-2023.7.22-py3-none-any.whl.metadata (2.2 kB)
Collecting mpmath>=0.19 (from sympy->torch==2.0.1)
  Using cached mpmath-1.3.0-py3-none-any.whl (536 kB)
Using cached Pillow-10.0.0-cp310-cp310-macosx_11_0_arm64.whl (3.1 MB)
Using cached filelock-3.12.3-py3-none-any.whl (11 kB)
Using cached typing_extensions-4.7.1-py3-none-any.whl (33 kB)
Using cached numpy-1.25.2-cp310-cp310-macosx_11_0_arm64.whl (14.0 MB)
Using cached requests-2.31.0-py3-none-any.whl (62 kB)
Using cached certifi-2023.7.22-py3-none-any.whl (158 kB)
Using cached charset_normalizer-3.2.0-cp310-cp310-macosx_11_0_arm64.whl (124 kB)
Using cached MarkupSafe-2.1.3-cp310-cp310-macosx_10_9_universal2.whl (17 kB)
Using cached urllib3-2.0.4-py3-none-any.whl (123 kB)
Installing collected packages: mpmath, urllib3, typing-extensions, sympy, pillow, numpy, networkx, MarkupSafe, idna, charset-normalizer, certifi, requests, jinja2, filelock, torch, torchvision
Successfully installed MarkupSafe-2.1.3 certifi-2023.7.22 charset-normalizer-3.2.0 filelock-3.12.3 idna-3.4 jinja2-3.1.2 mpmath-1.3.0 networkx-3.1 numpy-1.25.2 pillow-10.0.0 requests-2.31.0 sympy-1.12 torch-2.0.1 torchvision-0.15.2 typing-extensions-4.7.1 urllib3-2.0.4
Installing clip
Installing open_clip
Installing requirements for CodeFormer
Installing requirements
Launching Web UI with arguments: --skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
2023-09-01 18:18:30.799 Python[13219:522956] apply_selection_policy_once: avoid use of removable GPUs (via org.python.python:GPUSelectionPolicy->avoidRemovable)
Warning: caught exception 'Torch not compiled with CUDA enabled', memory monitor disabled
Loading weights [e948ca5dc4] from /Users/nick/PycharmProjects/stable-diffusion-webui/models/Stable-diffusion/absolutereality_v16.safetensors
Creating model from config: /Users/nick/PycharmProjects/stable-diffusion-webui/configs/v1-inference.yaml
Running on local URL:  http://127.0.0.1:7860

To create a public link, set `share=True` in `launch()`.
Startup time: 85.1s (prepare environment: 55.7s, import torch: 5.9s, import gradio: 5.6s, setup paths: 4.7s, initialize shared: 0.2s, other imports: 12.1s, setup codeformer: 0.1s, load scripts: 0.3s, initialize extra networks: 0.1s, create ui: 0.2s, gradio launch: 0.1s).
loading stable diffusion model: TypeError
Traceback (most recent call last):
  File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 973, in _bootstrap
    self._bootstrap_inner()
  File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 953, in run
    self._target(*self._args, **self._kwargs)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/initialize.py", line 147, in load_model
    shared.sd_model  # noqa: B018
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/shared_items.py", line 110, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 499, in get_sd_model
    load_model()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 626, in load_model
    load_model_weights(sd_model, checkpoint_info, state_dict, timer)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 353, in load_model_weights
    model.load_state_dict(state_dict, strict=False)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in <lambda>
    module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs))
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 221, in load_state_dict
    original(module, state_dict, strict=strict)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2027, in load_state_dict
    load(self, state_dict)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  [Previous line repeated 3 more times]
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2009, in load
    module._load_from_state_dict(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 226, in <lambda>
    conv2d_load_from_state_dict = self.replace(torch.nn.Conv2d, '_load_from_state_dict', lambda *args, **kwargs: load_from_state_dict(conv2d_load_from_state_dict, *args, **kwargs))
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 191, in load_from_state_dict
    module._parameters[name] = torch.nn.parameter.Parameter(torch.zeros_like(param, device=device, dtype=dtype), requires_grad=param.requires_grad)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1780, in zeros_like
    return aten.empty_like.default(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_ops.py", line 287, in __call__
    return self._op(*args, **kwargs or {})
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_refs/__init__.py", line 4254, in empty_like
    return torch.empty_strided(
TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.


Stable diffusion model failed to load
Applying attention optimization: sub-quadratic... done.
Loading weights [e948ca5dc4] from /Users/nick/PycharmProjects/stable-diffusion-webui/models/Stable-diffusion/absolutereality_v16.safetensors
Creating model from config: /Users/nick/PycharmProjects/stable-diffusion-webui/configs/v1-inference.yaml
loading stable diffusion model: TypeError
Traceback (most recent call last):
  File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 973, in _bootstrap
    self._bootstrap_inner()
  File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui_extra_networks.py", line 392, in pages_html
    return refresh()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui_extra_networks.py", line 398, in refresh
    pg.refresh()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui_extra_networks_textual_inversion.py", line 13, in refresh
    sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 255, in load_textual_inversion_embeddings
    self.expected_shape = self.get_expected_shape()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 154, in get_expected_shape
    vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/shared_items.py", line 110, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 499, in get_sd_model
    load_model()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 626, in load_model
    load_model_weights(sd_model, checkpoint_info, state_dict, timer)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 353, in load_model_weights
    model.load_state_dict(state_dict, strict=False)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in <lambda>
    module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs))
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 221, in load_state_dict
    original(module, state_dict, strict=strict)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2027, in load_state_dict
    load(self, state_dict)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  [Previous line repeated 3 more times]
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2009, in load
    module._load_from_state_dict(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 226, in <lambda>
    conv2d_load_from_state_dict = self.replace(torch.nn.Conv2d, '_load_from_state_dict', lambda *args, **kwargs: load_from_state_dict(conv2d_load_from_state_dict, *args, **kwargs))
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 191, in load_from_state_dict
    module._parameters[name] = torch.nn.parameter.Parameter(torch.zeros_like(param, device=device, dtype=dtype), requires_grad=param.requires_grad)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1780, in zeros_like
    return aten.empty_like.default(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_ops.py", line 287, in __call__
    return self._op(*args, **kwargs or {})
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_refs/__init__.py", line 4254, in empty_like
    return torch.empty_strided(
TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.


Stable diffusion model failed to load
Loading weights [e948ca5dc4] from /Users/nick/PycharmProjects/stable-diffusion-webui/models/Stable-diffusion/absolutereality_v16.safetensors
Creating model from config: /Users/nick/PycharmProjects/stable-diffusion-webui/configs/v1-inference.yaml
Traceback (most recent call last):
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/routes.py", line 488, in run_predict
    output = await app.get_blocks().process_api(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/blocks.py", line 1431, in process_api
    result = await self.call_function(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/blocks.py", line 1103, in call_function
    prediction = await anyio.to_thread.run_sync(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/to_thread.py", line 33, in run_sync
    return await get_asynclib().run_sync_in_worker_thread(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
    return await future
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui_extra_networks.py", line 392, in pages_html
    return refresh()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui_extra_networks.py", line 398, in refresh
    pg.refresh()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui_extra_networks_textual_inversion.py", line 13, in refresh
    sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 255, in load_textual_inversion_embeddings
    self.expected_shape = self.get_expected_shape()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 154, in get_expected_shape
    vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
AttributeError: 'NoneType' object has no attribute 'cond_stage_model'
loading stable diffusion model: TypeError
Traceback (most recent call last):
  File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 973, in _bootstrap
    self._bootstrap_inner()
  File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui_extra_networks.py", line 392, in pages_html
    return refresh()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui_extra_networks.py", line 398, in refresh
    pg.refresh()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui_extra_networks_textual_inversion.py", line 13, in refresh
    sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 255, in load_textual_inversion_embeddings
    self.expected_shape = self.get_expected_shape()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 154, in get_expected_shape
    vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/shared_items.py", line 110, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 499, in get_sd_model
    load_model()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 626, in load_model
    load_model_weights(sd_model, checkpoint_info, state_dict, timer)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 353, in load_model_weights
    model.load_state_dict(state_dict, strict=False)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in <lambda>
    module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs))
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 221, in load_state_dict
    original(module, state_dict, strict=strict)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2027, in load_state_dict
    load(self, state_dict)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  [Previous line repeated 3 more times]
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2009, in load
    module._load_from_state_dict(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 226, in <lambda>
    conv2d_load_from_state_dict = self.replace(torch.nn.Conv2d, '_load_from_state_dict', lambda *args, **kwargs: load_from_state_dict(conv2d_load_from_state_dict, *args, **kwargs))
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 191, in load_from_state_dict
    module._parameters[name] = torch.nn.parameter.Parameter(torch.zeros_like(param, device=device, dtype=dtype), requires_grad=param.requires_grad)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1780, in zeros_like
    return aten.empty_like.default(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_ops.py", line 287, in __call__
    return self._op(*args, **kwargs or {})
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_refs/__init__.py", line 4254, in empty_like
    return torch.empty_strided(
TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.


Stable diffusion model failed to load
Loading weights [e948ca5dc4] from /Users/nick/PycharmProjects/stable-diffusion-webui/models/Stable-diffusion/absolutereality_v16.safetensors
Traceback (most recent call last):
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/routes.py", line 488, in run_predict
    output = await app.get_blocks().process_api(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/blocks.py", line 1431, in process_api
    result = await self.call_function(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/blocks.py", line 1103, in call_function
    prediction = await anyio.to_thread.run_sync(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/to_thread.py", line 33, in run_sync
    return await get_asynclib().run_sync_in_worker_thread(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
    return await future
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui_extra_networks.py", line 392, in pages_html
    return refresh()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui_extra_networks.py", line 398, in refresh
    pg.refresh()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui_extra_networks_textual_inversion.py", line 13, in refresh
    sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 255, in load_textual_inversion_embeddings
    self.expected_shape = self.get_expected_shape()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 154, in get_expected_shape
    vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
AttributeError: 'NoneType' object has no attribute 'cond_stage_model'
Creating model from config: /Users/nick/PycharmProjects/stable-diffusion-webui/configs/v1-inference.yaml
loading stable diffusion model: TypeError
Traceback (most recent call last):
  File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 973, in _bootstrap
    self._bootstrap_inner()
  File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui.py", line 1298, in <lambda>
    update_image_cfg_scale_visibility = lambda: gr.update(visible=shared.sd_model and shared.sd_model.cond_stage_key == "edit")
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/shared_items.py", line 110, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 499, in get_sd_model
    load_model()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 626, in load_model
    load_model_weights(sd_model, checkpoint_info, state_dict, timer)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 353, in load_model_weights
    model.load_state_dict(state_dict, strict=False)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in <lambda>
    module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs))
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 221, in load_state_dict
    original(module, state_dict, strict=strict)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2027, in load_state_dict
    load(self, state_dict)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  [Previous line repeated 3 more times]
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2009, in load
    module._load_from_state_dict(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 226, in <lambda>
    conv2d_load_from_state_dict = self.replace(torch.nn.Conv2d, '_load_from_state_dict', lambda *args, **kwargs: load_from_state_dict(conv2d_load_from_state_dict, *args, **kwargs))
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 191, in load_from_state_dict
    module._parameters[name] = torch.nn.parameter.Parameter(torch.zeros_like(param, device=device, dtype=dtype), requires_grad=param.requires_grad)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1780, in zeros_like
    return aten.empty_like.default(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_ops.py", line 287, in __call__
    return self._op(*args, **kwargs or {})
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_refs/__init__.py", line 4254, in empty_like
    return torch.empty_strided(
TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.


Stable diffusion model failed to load
Loading weights [e948ca5dc4] from /Users/nick/PycharmProjects/stable-diffusion-webui/models/Stable-diffusion/absolutereality_v16.safetensors
Creating model from config: /Users/nick/PycharmProjects/stable-diffusion-webui/configs/v1-inference.yaml
loading stable diffusion model: TypeError
Traceback (most recent call last):
  File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 973, in _bootstrap
    self._bootstrap_inner()
  File "/opt/homebrew/Cellar/python@3.10/3.10.13/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/ui.py", line 1298, in <lambda>
    update_image_cfg_scale_visibility = lambda: gr.update(visible=shared.sd_model and shared.sd_model.cond_stage_key == "edit")
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/shared_items.py", line 110, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 499, in get_sd_model
    load_model()
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 626, in load_model
    load_model_weights(sd_model, checkpoint_info, state_dict, timer)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_models.py", line 353, in load_model_weights
    model.load_state_dict(state_dict, strict=False)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in <lambda>
    module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs))
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 221, in load_state_dict
    original(module, state_dict, strict=strict)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2027, in load_state_dict
    load(self, state_dict)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  [Previous line repeated 3 more times]
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2009, in load
    module._load_from_state_dict(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 226, in <lambda>
    conv2d_load_from_state_dict = self.replace(torch.nn.Conv2d, '_load_from_state_dict', lambda *args, **kwargs: load_from_state_dict(conv2d_load_from_state_dict, *args, **kwargs))
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/modules/sd_disable_initialization.py", line 191, in load_from_state_dict
    module._parameters[name] = torch.nn.parameter.Parameter(torch.zeros_like(param, device=device, dtype=dtype), requires_grad=param.requires_grad)
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1780, in zeros_like
    return aten.empty_like.default(
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_ops.py", line 287, in __call__
    return self._op(*args, **kwargs or {})
  File "/Users/nick/PycharmProjects/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_refs/__init__.py", line 4254, in empty_like
    return torch.empty_strided(
TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.


Stable diffusion model failed to load

@akx
Copy link
Collaborator

akx commented Sep 1, 2023

Perhaps core packages such as threading.py are not copied into the virtual environment when a full path including major.minor.patch (3.10.13) version of Python is used?

They should be, and that shouldn't be the issue.

Can you rename your ui-config.json to e.g. ui-config.json.backup and try again, just in case this is an issue with some botched setting? Also, can you try with another model than absolutereality_v16.safetensors?

@0-vortex
Copy link

0-vortex commented Sep 1, 2023

Perhaps core packages such as threading.py are not copied into the virtual environment when a full path including major.minor.patch (3.10.13) version of Python is used?

They should be, and that shouldn't be the issue.

Can you rename your ui-config.json to e.g. ui-config.json.backup and try again, just in case this is an issue with some botched setting? Also, can you try with another model than absolutereality_v16.safetensors?

I tried that, and new clean install even:

➜ ./webui.sh

################################################################
Install script for stable-diffusion + Web UI
Tested on Debian 11 (Bullseye)
################################################################

################################################################
Running on vortex user
################################################################

################################################################
Repo already cloned, using it as install directory
################################################################

################################################################
Create and activate python venv
################################################################

################################################################
Launching launch.py...
################################################################
Python 3.10.4 (v3.10.4:9d38120e33, Mar 23 2022, 17:29:05) [Clang 13.0.0 (clang-1300.0.29.30)]
Version: v1.6.0
Commit hash: 5ef669de080814067961f28357256e8fe27544f4
Launching Web UI with arguments: --skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
Warning: caught exception 'Torch not compiled with CUDA enabled', memory monitor disabled
*** "Disable all extensions" option was set, will not load any extensions ***
Loading weights [4199bcdd14] from /Users/vortex/WebstormProjectsAi/stable-diffusion-webui/models/Stable-diffusion/ani/revAnimated_v122.safetensors
Creating model from config: /Users/vortex/WebstormProjectsAi/stable-diffusion-webui/configs/v1-inference.yaml
Running on local URL:  http://127.0.0.1:7860

To create a public link, set `share=True` in `launch()`.
Startup time: 3.8s (import torch: 1.1s, import gradio: 0.7s, setup paths: 0.3s, initialize shared: 0.2s, other imports: 0.4s, scripts list_optimizers: 0.3s, initialize extra networks: 0.1s, create ui: 0.2s, gradio launch: 0.3s).
loading stable diffusion model: TypeError
Traceback (most recent call last):
  File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 966, in _bootstrap
    self._bootstrap_inner()
  File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 1009, in _bootstrap_inner
    self.run()
  File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 946, in run
    self._target(*self._args, **self._kwargs)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/initialize.py", line 147, in load_model
    shared.sd_model  # noqa: B018
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/shared_items.py", line 110, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_models.py", line 499, in get_sd_model
    load_model()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_models.py", line 626, in load_model
    load_model_weights(sd_model, checkpoint_info, state_dict, timer)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_models.py", line 353, in load_model_weights
    model.load_state_dict(state_dict, strict=False)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in <lambda>
    module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs))
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 221, in load_state_dict
    original(module, state_dict, strict=strict)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2027, in load_state_dict
    load(self, state_dict)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  [Previous line repeated 3 more times]
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2009, in load
    module._load_from_state_dict(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 226, in <lambda>
    conv2d_load_from_state_dict = self.replace(torch.nn.Conv2d, '_load_from_state_dict', lambda *args, **kwargs: load_from_state_dict(conv2d_load_from_state_dict, *args, **kwargs))
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 191, in load_from_state_dict
    module._parameters[name] = torch.nn.parameter.Parameter(torch.zeros_like(param, device=device, dtype=dtype), requires_grad=param.requires_grad)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1780, in zeros_like
    return aten.empty_like.default(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_ops.py", line 287, in __call__
    return self._op(*args, **kwargs or {})
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_refs/__init__.py", line 4254, in empty_like
    return torch.empty_strided(
TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.


Stable diffusion model failed to load
Applying attention optimization: sub-quadratic... done.
Loading weights [4199bcdd14] from /Users/vortex/WebstormProjectsAi/stable-diffusion-webui/models/Stable-diffusion/ani/revAnimated_v122.safetensors
Creating model from config: /Users/vortex/WebstormProjectsAi/stable-diffusion-webui/configs/v1-inference.yaml
loading stable diffusion model: TypeError
Traceback (most recent call last):
  File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 966, in _bootstrap
    self._bootstrap_inner()
  File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 1009, in _bootstrap_inner
    self.run()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 867, in run
    result = context.run(func, *args)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui_extra_networks.py", line 392, in pages_html
    return refresh()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui_extra_networks.py", line 398, in refresh
    pg.refresh()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui_extra_networks_textual_inversion.py", line 13, in refresh
    sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 255, in load_textual_inversion_embeddings
    self.expected_shape = self.get_expected_shape()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 154, in get_expected_shape
    vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/shared_items.py", line 110, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_models.py", line 499, in get_sd_model
    load_model()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_models.py", line 626, in load_model
    load_model_weights(sd_model, checkpoint_info, state_dict, timer)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_models.py", line 353, in load_model_weights
    model.load_state_dict(state_dict, strict=False)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in <lambda>
    module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs))
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 221, in load_state_dict
    original(module, state_dict, strict=strict)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2027, in load_state_dict
    load(self, state_dict)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  [Previous line repeated 3 more times]
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2009, in load
    module._load_from_state_dict(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 226, in <lambda>
    conv2d_load_from_state_dict = self.replace(torch.nn.Conv2d, '_load_from_state_dict', lambda *args, **kwargs: load_from_state_dict(conv2d_load_from_state_dict, *args, **kwargs))
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 191, in load_from_state_dict
    module._parameters[name] = torch.nn.parameter.Parameter(torch.zeros_like(param, device=device, dtype=dtype), requires_grad=param.requires_grad)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1780, in zeros_like
    return aten.empty_like.default(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_ops.py", line 287, in __call__
    return self._op(*args, **kwargs or {})
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_refs/__init__.py", line 4254, in empty_like
    return torch.empty_strided(
TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.


Stable diffusion model failed to load
Loading weights [4199bcdd14] from /Users/vortex/WebstormProjectsAi/stable-diffusion-webui/models/Stable-diffusion/ani/revAnimated_v122.safetensors
Creating model from config: /Users/vortex/WebstormProjectsAi/stable-diffusion-webui/configs/v1-inference.yaml
Traceback (most recent call last):
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/routes.py", line 488, in run_predict
    output = await app.get_blocks().process_api(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/blocks.py", line 1431, in process_api
    result = await self.call_function(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/blocks.py", line 1103, in call_function
    prediction = await anyio.to_thread.run_sync(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/to_thread.py", line 31, in run_sync
    return await get_asynclib().run_sync_in_worker_thread(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
    return await future
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 867, in run
    result = context.run(func, *args)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui_extra_networks.py", line 392, in pages_html
    return refresh()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui_extra_networks.py", line 398, in refresh
    pg.refresh()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui_extra_networks_textual_inversion.py", line 13, in refresh
    sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 255, in load_textual_inversion_embeddings
    self.expected_shape = self.get_expected_shape()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 154, in get_expected_shape
    vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
AttributeError: 'NoneType' object has no attribute 'cond_stage_model'
loading stable diffusion model: TypeError
Traceback (most recent call last):
  File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 966, in _bootstrap
    self._bootstrap_inner()
  File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 1009, in _bootstrap_inner
    self.run()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 867, in run
    result = context.run(func, *args)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui_extra_networks.py", line 392, in pages_html
    return refresh()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui_extra_networks.py", line 398, in refresh
    pg.refresh()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui_extra_networks_textual_inversion.py", line 13, in refresh
    sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 255, in load_textual_inversion_embeddings
    self.expected_shape = self.get_expected_shape()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 154, in get_expected_shape
    vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/shared_items.py", line 110, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_models.py", line 499, in get_sd_model
    load_model()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_models.py", line 626, in load_model
    load_model_weights(sd_model, checkpoint_info, state_dict, timer)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_models.py", line 353, in load_model_weights
    model.load_state_dict(state_dict, strict=False)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in <lambda>
    module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs))
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 221, in load_state_dict
    original(module, state_dict, strict=strict)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2027, in load_state_dict
    load(self, state_dict)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  [Previous line repeated 3 more times]
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2009, in load
    module._load_from_state_dict(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 226, in <lambda>
    conv2d_load_from_state_dict = self.replace(torch.nn.Conv2d, '_load_from_state_dict', lambda *args, **kwargs: load_from_state_dict(conv2d_load_from_state_dict, *args, **kwargs))
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 191, in load_from_state_dict
    module._parameters[name] = torch.nn.parameter.Parameter(torch.zeros_like(param, device=device, dtype=dtype), requires_grad=param.requires_grad)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1780, in zeros_like
    return aten.empty_like.default(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_ops.py", line 287, in __call__
    return self._op(*args, **kwargs or {})
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_refs/__init__.py", line 4254, in empty_like
    return torch.empty_strided(
TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.


Stable diffusion model failed to load
Loading weights [4199bcdd14] from /Users/vortex/WebstormProjectsAi/stable-diffusion-webui/models/Stable-diffusion/ani/revAnimated_v122.safetensors
Traceback (most recent call last):
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/routes.py", line 488, in run_predict
    output = await app.get_blocks().process_api(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/blocks.py", line 1431, in process_api
    result = await self.call_function(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/blocks.py", line 1103, in call_function
    prediction = await anyio.to_thread.run_sync(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/to_thread.py", line 31, in run_sync
    return await get_asynclib().run_sync_in_worker_thread(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
    return await future
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 867, in run
    result = context.run(func, *args)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui_extra_networks.py", line 392, in pages_html
    return refresh()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui_extra_networks.py", line 398, in refresh
    pg.refresh()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui_extra_networks_textual_inversion.py", line 13, in refresh
    sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
Creating model from config: /Users/vortex/WebstormProjectsAi/stable-diffusion-webui/configs/v1-inference.yaml
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 255, in load_textual_inversion_embeddings
    self.expected_shape = self.get_expected_shape()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/textual_inversion/textual_inversion.py", line 154, in get_expected_shape
    vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
AttributeError: 'NoneType' object has no attribute 'cond_stage_model'
loading stable diffusion model: TypeError
Traceback (most recent call last):
  File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 966, in _bootstrap
    self._bootstrap_inner()
  File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/threading.py", line 1009, in _bootstrap_inner
    self.run()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 867, in run
    result = context.run(func, *args)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/ui.py", line 1298, in <lambda>
    update_image_cfg_scale_visibility = lambda: gr.update(visible=shared.sd_model and shared.sd_model.cond_stage_key == "edit")
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/shared_items.py", line 110, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_models.py", line 499, in get_sd_model
    load_model()
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_models.py", line 626, in load_model
    load_model_weights(sd_model, checkpoint_info, state_dict, timer)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_models.py", line 353, in load_model_weights
    model.load_state_dict(state_dict, strict=False)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in <lambda>
    module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs))
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 221, in load_state_dict
    original(module, state_dict, strict=strict)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2027, in load_state_dict
    load(self, state_dict)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2015, in load
    load(child, child_state_dict, child_prefix)
  [Previous line repeated 3 more times]
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2009, in load
    module._load_from_state_dict(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 226, in <lambda>
    conv2d_load_from_state_dict = self.replace(torch.nn.Conv2d, '_load_from_state_dict', lambda *args, **kwargs: load_from_state_dict(conv2d_load_from_state_dict, *args, **kwargs))
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/modules/sd_disable_initialization.py", line 191, in load_from_state_dict
    module._parameters[name] = torch.nn.parameter.Parameter(torch.zeros_like(param, device=device, dtype=dtype), requires_grad=param.requires_grad)
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1780, in zeros_like
    return aten.empty_like.default(
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_ops.py", line 287, in __call__
    return self._op(*args, **kwargs or {})
  File "/Users/vortex/WebstormProjectsAi/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/_refs/__init__.py", line 4254, in empty_like
    return torch.empty_strided(
TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.


Stable diffusion model failed to load

@0-vortex
Copy link

0-vortex commented Sep 1, 2023

Forgot to mention that even with the errors, in the second instance basic generating stuff works but when trying hi-res it goes out of memory like

loc("MM/(batch1@batch2)"("(mpsFileLoc): /AppleInternal/Library/BuildRoots/0783246a-4091-11ee-8fca-aead88ae2785/Library/Caches/com.apple.xbs/Sources/MetalPerformanceShadersGraph/mpsgraph/MetalPerformanceShadersGraph/Core/Files/MPSGraphUtilities.mm":39:0)): error: 'anec.matmul' op Invalid configuration for the following reasons: Tensor dimensions N8D1C512H1W38400 are not within supported range, N[1-65536]D[1-16384]C[1-65536]H[1-16384]W[1-16384].

When it previously wouldn't, that's a 512x768 image I upscaled last week :<

@nicklansley
Copy link

@akx yes that worked! Renaming the ui-config.json to ui-config.json.backup fixed the issue and now it is working fine including using the GPU (and the Macbook fans!). See screenshot of Activity monitor - GPU at 95.8% on first line (Python):
Screenshot 2023-09-01 at 20 20 43

@akx
Copy link
Collaborator

akx commented Sep 1, 2023

@nicklansley Great to hear! It'd be nice to see the backup and the newly generated config so we can figure out what went wrong from the diff. If you don't feel like sharing the files here, you can send them as attachments to (my github username) at iki dot fi.

@nicklansley
Copy link

@akx - I have just performed a 'diff ui-config.json ui-config.json.backup' and there are no differences!

I am still using absolutereality_v16.safetensors and can change to other models and they work too. So, the act of the application not being able to find ui-config.json and having to go through the process of rebuilding it may fix some issue?

Anyway, here is my ui-config.json that works, not forgetting is identical to ui-config.json.backup
I have added a .txt extension because of GitHub restrictions:

ui-config.json.txt

@akx
Copy link
Collaborator

akx commented Sep 1, 2023

@nicklansley Did you by any chance rename he file while webui was running? I'm trying to think of how the files could be identical and still have one work and the other not...

@yanis-git
Copy link

I jump to your discussion :) i am facing same issue as nicklansley mentioned, I have followed your instruction and ensure to use venv only. Stacktrace is very similar with the final message.

TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.

I notice that if i switch between two models from the UI dropdown, then it is working

@nicklansley
Copy link

@akx I have done some further testing to understand why renaming the file seemed to solve the problem. I found out that the issue is not related to the file name, but to the first model that the application loads from the models/Stable-diffusion directory.

When I started the application with webui.sh, it automatically loaded the first model in alphabetical order, which was absolutereality_v16.safetensors. This model caused a stream of errors in the terminal, saying that it could not convert a MPS Tensor to float64 dtype (TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead. Stable diffusion model failed to load).

However, when I renamed this model to absolutereality_v16.safetensors.bak and restarted webui.sh, it loaded the next model in alphabetical order, which was epicrealism_pureEvolutionV3.safetensors. This model worked fine and did not cause any errors.

I then renamed absolutereality_v16.safetensors.bak back to absolutereality_v16.safetensors and restarted webui.sh again. This time, it did not load this model by default, but remembered the last model I used, which was epicrealism_pureEvolutionV3.safetensors. This model continued to worked fine.

I then switched to absolutereality_v16.safetensors from the UI, without restarting anything, and it also worked fine. No errors were shown in the terminal, demonstrating that there does not seem to be anything wrong with this model to cause any specific error.

However, when I restarted webui.sh again with absolutereality_v16.safetensors as the currently chosen model, the errors reappeared.

I also followed @yanis-git's suggestion and switched between all the different models from the UI, without restarting anything, and they all worked fine.

So, there is a bug in loading the first model on startup, even if that model is compatible with the application. Of course, there may be some specific issue with absolutereality_v16.safetensors but the crucial point is that it has been working fine up until this commit. Has application code changed to perform extra tests to decide the compatibility of models that this model fails on startup? For example, I assume you must detect the SD version (1.4, 1.5, 2.x, XL) of the model and that absolutereality_v16.safetensors is misrepresenting itself in some way? Just some thoughts...

If you want to try the model yourself (Absolute Reality v1.6) you can download it from the CivitAI catalogue here: https://civitai.com/models/81458?modelVersionId=108576 with a note that it has been superseded by v1.8 which does not have this issue as the startup model.

@0-vortex
Copy link

0-vortex commented Sep 2, 2023

@nicklansley @akx IMHO don't think the reasoning for blaming a checkpoint is correct. In my installations, where I got rid of the errors by just cycling, I never had version 1.6.0 of absolute reality (raw/absolutereality_v181.safetensors in my logs). Furthermore, the MPS errors persist when trying to do hi-res, can't even hi-res a 512x512 on an M1 max, takes 20s/it, while in the 1.5.2 folder that's 6-7s/it. There are breaking changes to the RNG, can't reproduce an image with the same config.json, even tried diffing the 1.5.2 config with the 1.6.0 config to make sure I have everything set up the same way (some settings changed name) and couldn't figure it out. The closest I can get the images to match my previous generations is with the NV RNG, but they still are not the same and the colours are a bit off.

If put to guess, hi-res making 2 passes, how is memory managed there?

@nicklansley
Copy link

@0-vortex thanks for the update - I seem to be running hires OK with the application using just over 6GB of memory for 512x512 then switching up to just over 10GB for hires 1024x1024 (according to Activity Monitor).

If you run Activity Monitor and display the GPU History window when running the app, does this window show the GPU maxxing out during generation, or not used at all? When I followed the advice from @ngtongsheng it worked but only used the CPU and I was getting the same seconds per iteration performance that you experienced (surprisingly good for CPU - all hail RISC architecture!) but way to slow for normal use. So just check your GPU is being engaged.

@0-vortex
Copy link

0-vortex commented Sep 2, 2023

@0-vortex thanks for the update - I seem to be running hires OK with the application using just over 6GB of memory for 512x512 then switching up to just over 10GB for hires 1024x1024 (according to Activity Monitor).

If you run Activity Monitor and display the GPU History window when running the app, does this window show the GPU maxxing out during generation, or not used at all? When I followed the advice from @ngtongsheng it worked but only used the CPU and I was getting the same seconds per iteration performance that you experienced (surprisingly good for CPU - all hail RISC architecture!) but way to slow for normal use. So just check your GPU is being engaged.

It spikes to 75-80% just generating the 512x512 image for me, 95%+ on hi-res :<

@soarcreator
Copy link

I have the issue that seems the same as this on my Intel mac with GPU, and it seems no effect to rename ui-config.json to ui-config.json.backup. I don't know much about stable diffusion because I just started using stable diffusion today, but the last line says "TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead."
"--skip-torch-cuda-test --no-half --use-cpu all" solves my problem, but it takes too long to complete. Hope it will be fixed soon.

@ericwagner101
Copy link

I had the same issue and at least for me it was related to a messed up virtualenv. I had to do this:

  1. Clone a new copy of a1111
  2. cd to the new cloned directory.
  3. Run
    virtualenv venv -p python3
    source venv/bin/activate
  4. Then in the same shell window run webui.sh and let it repull all the dependencies.

After this the issue with certain older models went away.

@ngtongsheng
Copy link

ngtongsheng commented Sep 3, 2023

I had the same issue and at least for me it was related to a messed up virtualenv. I had to do this:

  1. Clone a new copy of a1111
  2. cd to the new cloned directory.
  3. Run
    virtualenv venv -p python3
    source venv/bin/activate
  4. Then in the same shell window run webui.sh and let it repull all the dependencies.

After this the issue with certain older models went away.

@ericwagner101 Thank you very much, tried on my macbook pro 13inch M1 chip and confirmed it works.

512x512 took 38secs, not too bad

@njulhy
Copy link

njulhy commented Sep 4, 2023

I faced the same issue, the following works for me

  • Delete webui-maco-env.sh
  • Remove export PYTORCH_MPS_HIGH_WATERMARK_RATIO="0.0" in webui-user.sh if exist
  • Add this env export COMMANDLINE_ARGS="--skip-torch-cuda-test --no-half --use-cpu all"

works for me.

@akx
Copy link
Collaborator

akx commented Sep 4, 2023

use-cpu all will make the webui not use MPS, which is quite certainly not what you want.

@brkirch
Copy link
Collaborator

brkirch commented Sep 6, 2023

I haven't been able to reproduce this issue, but #13099 will hopefully fix it. Please let me know if there are any further issues with that fix applied.

@ericwagner101
Copy link

ericwagner101 commented Sep 7, 2023

I haven't been able to reproduce this issue, but #13099 will hopefully fix it. Please let me know if there are any further issues with that fix applied.

Thanks. I applied the fix and tested it. It did change something as the original error does not occur, but there is a new ser of errors and the model still does not load. Snip of the errors:

File "/Users/eric/ai/stable-diffusion-webui/modules/sd_disable_initialization.py", line 225, in
module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs))
File "/Users/eric/ai/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in load_state_dict
original(module, state_dict, strict=strict)
File "/Users/eric/ai/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2041, in load_state_dict
raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format(
RuntimeError: Error(s) in loading state_dict for DiffusionEngine:
While copying the parameter named "model.diffusion_model.output_blocks.2.0.in_layers.0.weight", whose dimensions in the model are torch.Size([1920]) and whose dimensions in the checkpoint are torch.Size([1920]), an exception occurred : ('Cannot copy out of meta tensor; no data!',).
While copying the parameter named "model.diffusion_model.output_blocks.2.0.in_layers.0.bias", whose dimensions in the model are torch.Size([1920]) and whose dimensions in the checkpoint are torch.Size([1920]), an exception occurred : ('Cannot copy out of meta tensor; no data!',).
While copying the parameter named "model.diffusion_model.output_blocks.2.0.in_layers.2.weight", whose dimensions in the model are torch.Size([1280, 1920, 3, 3]) and whose dimensions in the checkpoint are torch.Size([1280, 1920, 3, 3]), an exception occurred : ('Cannot copy out of meta tensor; no data!',).

@dlnetworks
Copy link

dlnetworks commented Sep 8, 2023

I haven't been able to reproduce this issue, but #13099 will hopefully fix it. Please let me know if there are any further issues with that fix applied.

i was getting the same error after a git pull which brought me to 1.6.0.

im on a 2019 intel mac pro with an amd rx 6800 xt, python 3.10, torch 1.13 (but also works with torch 1.12)(still cant seem to get anything working with gpu using torch 2.x, or sdxl anywhere.. but thats outside the scope of this thread)

i first tried --disable-model-loading-ram-optimization when launching, while that did not produce the error, models were still not loading properly.. i was getting a bunch of stuff about producing NaN's and would only seem to create an all black image.

after making the changes in #13099, and omitting --disable-model-loading-ram-optimization from the launch command, im no longer seeing that error and so far i seem to be able to load models without any issues and generate images with gpu acceleration.

so.. nice one =]

one thing i did notice was apparently the new default for attention optimization is now sub-quadratic, as it was invokeai before the upgrade...

and i also noticed when switching to a different model, in the console it says:

Reusing loaded model to load

in any case.. thanks

@qlliu
Copy link

qlliu commented Sep 8, 2023

Tried to use --disable-model-loading-ram-optimization and can workaround this error.

Works for me

@jorge-campo
Copy link

Mac M1 here. I started to find this annoying issue about a week ago. None of the solutions in this thread worked for me. After multiple tests with clean A1111 installations, I'm convinced the issue is triggered by some checkpoints but not others. I can't find a reason for that, but in my case, tests are conclusive.

I pulled the #13099 PR and tested it, but it didn't make any difference, unfortunately.

@ericwagner101
Copy link

I haven't been able to reproduce this issue, but #13099 will hopefully fix it. Please let me know if there are any further issues with that fix applied.

Thanks. I applied the fix and tested it. It did change something as the original error does not occur, but there is a new ser of errors and the model still does not load. Snip of the errors:

File "/Users/eric/ai/stable-diffusion-webui/modules/sd_disable_initialization.py", line 225, in module_load_state_dict = self.replace(torch.nn.Module, 'load_state_dict', lambda *args, **kwargs: load_state_dict(module_load_state_dict, *args, **kwargs)) File "/Users/eric/ai/stable-diffusion-webui/modules/sd_disable_initialization.py", line 223, in load_state_dict original(module, state_dict, strict=strict) File "/Users/eric/ai/stable-diffusion-webui/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 2041, in load_state_dict raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format( RuntimeError: Error(s) in loading state_dict for DiffusionEngine: While copying the parameter named "model.diffusion_model.output_blocks.2.0.in_layers.0.weight", whose dimensions in the model are torch.Size([1920]) and whose dimensions in the checkpoint are torch.Size([1920]), an exception occurred : ('Cannot copy out of meta tensor; no data!',). While copying the parameter named "model.diffusion_model.output_blocks.2.0.in_layers.0.bias", whose dimensions in the model are torch.Size([1920]) and whose dimensions in the checkpoint are torch.Size([1920]), an exception occurred : ('Cannot copy out of meta tensor; no data!',). While copying the parameter named "model.diffusion_model.output_blocks.2.0.in_layers.2.weight", whose dimensions in the model are torch.Size([1280, 1920, 3, 3]) and whose dimensions in the checkpoint are torch.Size([1280, 1920, 3, 3]), an exception occurred : ('Cannot copy out of meta tensor; no data!',).

I tried the fix #13099 again and it's now working.

@kelvin-zhao
Copy link

The solutions doesn't seem to work for me, always when the last step of generation error occurs, I'm not sure how to do a clean reinstall, I deleted my webui folder and followed the instructions to install, how do i clean the global env, etc? But comfyui is running on the other end, will it kill comfyui?

@OSS-42
Copy link

OSS-42 commented Sep 14, 2023

had the same MPS issue with my Mac M2 16Gb and v1.6.
I had the issue with certain checkpoints too.

I did the clean reinstall steps.
Clone a new copy of a1111 cd to the new cloned directory. Run virtualenv venv -p python3 source venv/bin/activate Then in the same shell window run webui.sh and let it repull all the dependencies.

and it seems to be working fine with the checkpoints I had issue with.

I also tried to add the 2 lines in the dependency.py file, but this generated another set of errors.

EDIT : it worked the first time. After I closed the browser and restarted form the terminal, the error showed again.

@skfoo
Copy link

skfoo commented Sep 19, 2023

It seems the problem is related to not loading the configs/v1-inference.yaml for checkpoints. If at startup I get this error, I switch to the base SD 1.5 model and then it says it loads that config. Then I can switch to another model.

A related issue though is I can't use an inpainting model I make (following the steps on the wiki, where you merge three models). It tries loading the configs/v1-inference.yaml instead of v1-inpainting-inference. Selecting the SD 1.5 inpainting model does load the painting config, but when it doesn't maintain it when I select my own inpainting model, it reverts to the non-inpainting config.

I get the same problem when I switch to the SD XL 1.0 checkpoint and then switch to an SD 1.5; I have to switch to the base SD 1.5 model so it loads the v1-inference.yaml again and then select the checkpoint I want to use.

@DanBurkhardt
Copy link

It seems the problem is related to not loading the configs/v1-inference.yaml for checkpoints. If at startup I get this error, I switch to the base SD 1.5 model and then it says it loads that config. Then I can switch to another model.

that was the perfect solution for me-- so silly and simple. Thanks!

@cursed-github
Copy link

cursed-github commented Sep 28, 2023

I faced the same issue, the following works for me

  • Delete webui-maco-env.sh
  • Remove export PYTORCH_MPS_HIGH_WATERMARK_RATIO="0.0" in webui-user.sh if exist
  • Add this env export COMMANDLINE_ARGS="--skip-torch-cuda-test --no-half --use-cpu all"

I tried this and it works but now the whole thing is much slower than before, any ideas on how to solve this?

edit: I git reset to the previous version, and everything works fine and fast. Lesson learned - don't git pull anything that's not broken

@akx
Copy link
Collaborator

akx commented Oct 1, 2023

@cursed-github As said multiple times in this thread, --use-cpu all makes webui skip using the GPU at all, which is slow.

@ShinChven
Copy link

I had the same issue and temporarily resolved it by switching models.

@meegle84
Copy link

Tried to use --disable-model-loading-ram-optimization and can workaround this error.

working for me

@rilzi
Copy link

rilzi commented Oct 29, 2023

I had the same issue and at least for me it was related to a messed up virtualenv. I had to do this:我有同样的问题,至少对我来说,它与一个混乱的virtualenv有关。我必须这样做:

  1. Clone a new copy of a1111克隆 a1111 的新副本
  2. cd to the new cloned directory.cd 到新克隆的目录。
  3. Run 跑
    virtualenv venv -p python3
    source venv/bin/activate资料来源: venv/bin/activate
  4. Then in the same shell window run webui.sh and let it repull all the dependencies.然后,在同一个 shell 窗口中运行 webui.sh 并让它重新拉取所有依赖项。

After this the issue with certain older models went away.在此之后,某些旧型号的问题消失了。

can u show in video and how to fix this issue.plz

@eyes1993
Copy link

在设置里,勾选 ☑️将交叉关注层向上转换到 float32

@azflin
Copy link

azflin commented Nov 12, 2023

Tried to use --disable-model-loading-ram-optimization and can workaround this error.

BASED AND GOATED

@HAjiMa666
Copy link

I faced the same issue, the following works for me

  • Delete webui-maco-env.sh
  • Remove export PYTORCH_MPS_HIGH_WATERMARK_RATIO="0.0" in webui-user.sh if exist
  • Add this env export COMMANDLINE_ARGS="--skip-torch-cuda-test --no-half --use-cpu all"

works for me.

+1 it works fine

@YuuLinz
Copy link

YuuLinz commented Dec 2, 2023

我跳到你的讨论:)我面临着与nicklansley提到的同样的问题,我已遵循你的指示并确保仅使用 venv。Stacktrace 与最终消息非常相似。

TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.

我注意到,如果我从 UI 下拉列表中在两个模型之间切换,那么它就可以工作

yes 我遇到了同样的问题 但是在切换模型后能够正常使用,每次启动的第一个模型都会保存,我很难理解这是为什么

@andymove
Copy link

andymove commented Dec 5, 2023

I tried the solutions given

This works for me. 2 steps:

  1. check the option here: Settings->Stable Diffusion->Upcast cross attention layer to float32
  2. ./webui.sh --no-half

More Info:
When using other checkpoint, may still cause the error:
TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.

Work·around:
Switch Stable Diffusion checkpoint to 'v1-5-pruned-emaonly.ckpt [cc6cb27103]';
and then switch back.
The error disappear.

FYI. I got the solution base on this hint when I click 'Generate' button in the 'txt2img' tab :
NansException: A tensor with all NaNs was produced in Unet. This could be either because there's not enough precision to represent the picture, or because your video card does not support half type. Try setting the "Upcast cross attention layer to float32" option in Settings > Stable Diffusion or using the --no-half commandline argument to fix this. Use --disable-nan-check commandline argument to disable this check.

@luffy2025
Copy link

luffy2025 commented Dec 10, 2023

I also encountered the same issue, and I believe it was caused by loading the model during the initial startup. Here's how I resolved it:

  1. In the "models/Stable-diffusion" folder, I kept only one reliable "safetensors" file, such as "AnythingV5V3_v5PrtRE.safetensors" or "v1-5-pruned.safetensors."
  2. I restarted the ./webui.sh, and there were no error messages.
  3. I placed the other models back into their original folder.
  4. I started the ./webui.sh again, and there were still no error messages.
    The problem was resolved. My environment is M3 Max. Hope this can help you out.

@ShinChven
Copy link

I also encountered the same issue, and I believe it was caused by loading the model during the initial startup. Here's how I resolved it:

  1. In the "models/Stable-diffusion" folder, I kept only one reliable "safetensors" file, such as "AnythingV5V3_v5PrtRE.safetensors" or "v1-5-pruned.safetensors."
  2. I restarted the ./webui.sh, and there were no error messages.
  3. I placed the other models back into their original folder.
  4. I started the ./webui.sh again, and there were still no error messages.
    The problem was resolved. My environment is M3 Max. Hope this can help you out.

Switching the models can also solve this problem temporarily.

@esonhon
Copy link

esonhon commented Feb 16, 2024

我跳到你的讨论:)我面临着与nicklansley提到的同样的问题,我已遵循你的指示并确保仅使用 venv。Stacktrace 与最终消息非常相似。

TypeError: Cannot convert a MPS Tensor to float64 dtype as the MPS framework doesn't support float64. Please use float32 instead.

我注意到,如果我从 UI 下拉列表中在两个模型之间切换,那么它就可以工作

yes 我遇到了同样的问题 但是在切换模型后能够正常使用,每次启动的第一个模型都会保存,我很难理解这是为什么

Yes, it works for me.

@leemax
Copy link

leemax commented Mar 6, 2024

这是有些模型加载的时候,出现的错误,导致无法加载。
我的解决方案:
升级pytorh到2.2.1,会提示torchvision0.15.2对torch 2.0.1有依赖关系,把torchvision升级到0.17.1
现在可以成功加载chilloutrealistic_v21.safetensors模型

@andupotorac
Copy link

Tried to use --disable-model-loading-ram-optimization and can workaround this error.

BASED AND GOATED

Worked for me too. The error is gone.

@bytqq07
Copy link

bytqq07 commented May 18, 2024

I had the same issue and at least for me it was related to a messed up virtualenv. I had to do this:

  1. Clone a new copy of a1111
  2. cd to the new cloned directory.
  3. Run
    virtualenv venv -p python3
    source venv/bin/activate
  4. Then in the same shell window run webui.sh and let it repull all the dependencies.

After this the issue with certain older models went away.

hey, but if I get this error?

"source venv/bin/activate
zsh: command not found: virtualenv"

@DDxxgg
Copy link

DDxxgg commented Aug 5, 2024

--disable-model-loading-ram-optimization

it works for me too, thanks.

@ghost
Copy link

ghost commented Aug 8, 2024

I had the same issue and at least for me it was related to a messed up virtualenv. I had to do this:

  1. Clone a new copy of a1111
  2. cd to the new cloned directory.
  3. Run
    virtualenv venv -p python3
    source venv/bin/activate
  4. Then in the same shell window run webui.sh and let it repull all the dependencies.

After this the issue with certain older models went away.

Mey I ask what is a1111?

@smlbiobot
Copy link

@ukateki a1111 = Automatic1111 = this repo

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug-report Report of a bug, yet to be confirmed platform:mac Issues that apply to Apple OS X, M1, M2, etc
Projects
None yet
Development

Successfully merging a pull request may close this issue.