You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Stable diffusion model failed to load
Exception in thread MemMon:
Traceback (most recent call last):
File "C:\Users\xgevr\miniconda3\Lib\threading.py", line 1038, in _bootstrap_inner
self.run()
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\memmon.py", line 43, in run
torch.cuda.reset_peak_memory_stats()
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\cuda\memory.py", line 309, in reset_peak_memory_stats
return torch._C._cuda_resetPeakMemoryStats(device)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
Using already loaded model v1-5-pruned-emaonly.safetensors [6ce0161689]: done in 0.0s
RuntimeError: invalid argument to reset_peak_memory_stats
*** Error completing request
*** Arguments: ('task(jpo8trnumwzeiw5)', <gradio.routes.Request object at 0x0000021B1676FED0>, 'A dog', '', [], 20, 'DPM++ 2M Karras', 1, 1, 7, 512, 512, False, 0.7, 2, 'Latent', 0, 0, 0, 'Use same checkpoint', 'Use same sampler', '', '', [], 0, False, '', 0.8, -1, False, -1, 0, 0, 0, False, False, 'positive', 'comma', 0, False, False, 'start', '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, False, False, False, 0, False) {}
Traceback (most recent call last):
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\call_queue.py", line 57, in f
res = list(func(*args, **kwargs))
^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\call_queue.py", line 36, in f
res = func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\txt2img.py", line 110, in txt2img
processed = processing.process_images(p)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\processing.py", line 787, in process_images
res = process_images_inner(p)
^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\processing.py", line 940, in process_images_inner
model_hijack.embedding_db.load_textual_inversion_embeddings()
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\textual_inversion\textual_inversion.py", line 224, in load_textual_inversion_embeddings
self.expected_shape = self.get_expected_shape()
^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\textual_inversion\textual_inversion.py", line 156, in get_expected_shape
vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\sd_hijack_clip.py", line 344, in encode_embedding_init_text
embedded = embedding_layer.token_embedding.wrapped(ids.to(embedding_layer.token_embedding.wrapped.weight.device)).squeeze(0)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\nn\modules\sparse.py", line 163, in forward
return F.embedding(
^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\nn\functional.py", line 2237, in embedding
return torch.embedding(weight, input, padding_idx, scale_grad_by_freq, sparse)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: CUDA error: invalid argument
CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
For debugging consider passing CUDA_LAUNCH_BLOCKING=1.
Compile with TORCH_USE_CUDA_DSA to enable device-side assertions.
Traceback (most recent call last):
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\gradio\routes.py", line 488, in run_predict
output = await app.get_blocks().process_api(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\gradio\blocks.py", line 1431, in process_api
result = await self.call_function(
^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\gradio\blocks.py", line 1103, in call_function
prediction = await anyio.to_thread.run_sync(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\anyio\to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\anyio_backends_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\anyio_backends_asyncio.py", line 807, in run
result = context.run(func, *args)
^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\gradio\utils.py", line 707, in wrapper
response = f(args, **kwargs)
^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\call_queue.py", line 95, in f
mem_stats = {k: -(v//-(10241024)) for k, v in shared.mem_mon.stop().items()}
^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\memmon.py", line 99, in stop
return self.read()
^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\memmon.py", line 81, in read
torch_stats = torch.cuda.memory_stats(self.device)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\cuda\memory.py", line 258, in memory_stats
stats = memory_stats_as_nested_dict(device=device)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\cuda\memory.py", line 270, in memory_stats_as_nested_dict
return torch._C._cuda_memoryStats(device)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: invalid argument to memory_allocated
The text was updated successfully, but these errors were encountered:
Stable diffusion model failed to load
Exception in thread MemMon:
Traceback (most recent call last):
File "C:\Users\xgevr\miniconda3\Lib\threading.py", line 1038, in _bootstrap_inner
self.run()
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\memmon.py", line 43, in run
torch.cuda.reset_peak_memory_stats()
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\cuda\memory.py", line 309, in reset_peak_memory_stats
return torch._C._cuda_resetPeakMemoryStats(device)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
Using already loaded model v1-5-pruned-emaonly.safetensors [6ce0161689]: done in 0.0s
RuntimeError: invalid argument to reset_peak_memory_stats
*** Error completing request
*** Arguments: ('task(jpo8trnumwzeiw5)', <gradio.routes.Request object at 0x0000021B1676FED0>, 'A dog', '', [], 20, 'DPM++ 2M Karras', 1, 1, 7, 512, 512, False, 0.7, 2, 'Latent', 0, 0, 0, 'Use same checkpoint', 'Use same sampler', '', '', [], 0, False, '', 0.8, -1, False, -1, 0, 0, 0, False, False, 'positive', 'comma', 0, False, False, 'start', '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, False, False, False, 0, False) {}
Traceback (most recent call last):
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\call_queue.py", line 57, in f
res = list(func(*args, **kwargs))
^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\call_queue.py", line 36, in f
res = func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\txt2img.py", line 110, in txt2img
processed = processing.process_images(p)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\processing.py", line 787, in process_images
res = process_images_inner(p)
^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\processing.py", line 940, in process_images_inner
model_hijack.embedding_db.load_textual_inversion_embeddings()
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\textual_inversion\textual_inversion.py", line 224, in load_textual_inversion_embeddings
self.expected_shape = self.get_expected_shape()
^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\textual_inversion\textual_inversion.py", line 156, in get_expected_shape
vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\sd_hijack_clip.py", line 344, in encode_embedding_init_text
embedded = embedding_layer.token_embedding.wrapped(ids.to(embedding_layer.token_embedding.wrapped.weight.device)).squeeze(0)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\nn\modules\sparse.py", line 163, in forward
return F.embedding(
^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\nn\functional.py", line 2237, in embedding
return torch.embedding(weight, input, padding_idx, scale_grad_by_freq, sparse)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: CUDA error: invalid argument
CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
For debugging consider passing CUDA_LAUNCH_BLOCKING=1.
Compile with
TORCH_USE_CUDA_DSA
to enable device-side assertions.Traceback (most recent call last):
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\gradio\routes.py", line 488, in run_predict
output = await app.get_blocks().process_api(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\gradio\blocks.py", line 1431, in process_api
result = await self.call_function(
^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\gradio\blocks.py", line 1103, in call_function
prediction = await anyio.to_thread.run_sync(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\anyio\to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\anyio_backends_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\anyio_backends_asyncio.py", line 807, in run
result = context.run(func, *args)
^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\gradio\utils.py", line 707, in wrapper
response = f(args, **kwargs)
^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\call_queue.py", line 95, in f
mem_stats = {k: -(v//-(10241024)) for k, v in shared.mem_mon.stop().items()}
^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\memmon.py", line 99, in stop
return self.read()
^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\modules\memmon.py", line 81, in read
torch_stats = torch.cuda.memory_stats(self.device)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\cuda\memory.py", line 258, in memory_stats
stats = memory_stats_as_nested_dict(device=device)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\sd-test\Zluda\stable-diffusion-webui-directml\venv\Lib\site-packages\torch\cuda\memory.py", line 270, in memory_stats_as_nested_dict
return torch._C._cuda_memoryStats(device)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: invalid argument to memory_allocated
The text was updated successfully, but these errors were encountered: