Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Let's test the wheels #3

Open
woct0rdho opened this issue Oct 15, 2024 · 134 comments
Open

Let's test the wheels #3

woct0rdho opened this issue Oct 15, 2024 · 134 comments

Comments

@woct0rdho
Copy link
Owner

woct0rdho commented Oct 15, 2024

If you didn't, first read the instructions for installation here: https://github.com/woct0rdho/triton-windows#install-from-wheel

When you see errors, paste the whole error log, not only the last line

@woct0rdho
Copy link
Owner Author

woct0rdho commented Oct 15, 2024

@FurkanGozukara @shivshankar11 @osadchi @WingeD123 @frankyifei @jepjoo Please let me know if it works on your machine. You can use the simple test script here:
https://github.com/woct0rdho/triton?tab=readme-ov-file#test-if-it-works

@FurkanGozukara
Copy link

Awesome I saw. they are very small compared to older ones any particular reason?

I think I may test on CogVLM v2 what you think?

@woct0rdho
Copy link
Owner Author

Awesome I saw. they are very small compared to older ones any particular reason?

I excluded some binaries unnecessary for end users.

@WingeD123
Copy link

wow, thanks.
need CUDA/MSVC/Windows SDK, right?

@woct0rdho
Copy link
Owner Author

wow, thanks. need CUDA/MSVC/Windows SDK, right?

Yes, you can read the README again for more detailed instructions

@FurkanGozukara
Copy link

FurkanGozukara commented Oct 15, 2024

@woct0rdho test py failed for me

can't we make it include c++? it is super annoying and hard for people to install

I have got Build Tools version LTSC 17.8 and tested with Pytorch 2.4.1 and CUDA 12.4 and cudnn 8.9.7

user path
image

system path
image

image

here my c++ tools and SDKS

image

image

@shivshankar11
Copy link

even after setting all path and requirement triton compilation fail.

@woct0rdho
Copy link
Owner Author

@FurkanGozukara Please modify triton/runtime/build.py and try again. If it works, I'll publish new wheels

@@ -43,6 +43,11 @@
     # try to avoid setuptools if possible
     cc = os.environ.get("CC")
     if cc is None:
+        if os.name == "nt":
+            msvc_winsdk_inc_dirs, _ = find_msvc_winsdk()
+            if msvc_winsdk_inc_dirs:
+                cl_path = msvc_winsdk_inc_dirs[0].replace(r"\include", r"\bin\Hostx64\x64")
+            os.environ["PATH"] = cl_path + os.pathsep + os.environ["PATH"]
         # TODO: support more things here.
         cl = shutil.which("cl")
         gcc = shutil.which("gcc")

@FurkanGozukara
Copy link

FurkanGozukara commented Oct 15, 2024

@FurkanGozukara Please modify triton/runtime/build.py and try again. If it works, I'll publish new wheels

@@ -43,6 +43,11 @@
     # try to avoid setuptools if possible
     cc = os.environ.get("CC")
     if cc is None:
+        if os.name == "nt":
+            msvc_winsdk_inc_dirs, _ = find_msvc_winsdk()
+            if msvc_winsdk_inc_dirs:
+                cl_path = msvc_winsdk_inc_dirs[0].replace(r"\include", r"\bin\Hostx64\x64")
+            os.environ["PATH"] = cl_path + os.pathsep + os.environ["PATH"]
         # TODO: support more things here.
         cl = shutil.which("cl")
         gcc = shutil.which("gcc")

wait i used pre compiled wheel didnt compile on my system

@woct0rdho
Copy link
Owner Author

woct0rdho commented Oct 15, 2024

Modify this file in C:\Python310\Lib\site-packages\triton\runtime\build.py. The lines to modify are around Line 43

@FurkanGozukara
Copy link

FurkanGozukara commented Oct 15, 2024

@woct0rdho new error

(venv) C:\Users\Furkan\Pictures\ghgfhfg>python test.py
main.c
LINK : fatal error LNK1104: cannot open file 'python310.lib'
Traceback (most recent call last):
  File "C:\Users\Furkan\Pictures\ghgfhfg\test.py", line 26, in <module>
    b_compiled = add(a, a)
  File "C:\Users\Furkan\Pictures\ghgfhfg\test.py", line 21, in add
    add_kernel[grid](x, y, output, n_elements, BLOCK_SIZE=1024)
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\runtime\jit.py", line 345, in <lambda>
    return lambda *args, **kwargs: self.run(grid=grid, warmup=False, *args, **kwargs)
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\runtime\jit.py", line 607, in run
    device = driver.active.get_current_device()
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\runtime\driver.py", line 23, in __getattr__
    self._initialize_obj()
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\runtime\driver.py", line 20, in _initialize_obj
    self._obj = self._init_fn()
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\runtime\driver.py", line 9, in _create_driver
    return actives[0]()
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\backends\nvidia\driver.py", line 412, in __init__
    self.utils = CudaUtils()  # TODO: make static
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\backends\nvidia\driver.py", line 90, in __init__
    mod = compile_module_from_src(Path(os.path.join(dirname, "driver.c")).read_text(), "cuda_utils")
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\backends\nvidia\driver.py", line 67, in compile_module_from_src
    so = _build(name, src_path, tmpdir, library_dirs(), include_dir, libraries)
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\runtime\build.py", line 75, in _build
    ret = subprocess.check_call(cc_cmd)
  File "C:\Python310\lib\subprocess.py", line 369, in check_call
    raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.41.34120\\bin\\Hostx64\\x64\\cl.EXE', 'C:\\Users\\Furkan\\AppData\\Local\\Temp\\tmpob5k0vz6\\main.c', '/nologo', '/O2', '/LD', '/wd4819', '/IC:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\Lib\\site-packages\\triton\\backends\\nvidia\\include', '/IC:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.4\\include', '/IC:\\Users\\Furkan\\AppData\\Local\\Temp\\tmpob5k0vz6', '/IC:\\Python310\\Include', '/IC:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.41.34120\\include', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\shared', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\ucrt', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\um', '/link', '/LIBPATH:C:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\Lib\\site-packages\\triton\\backends\\nvidia\\lib', '/LIBPATH:C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.4\\lib\\x64', '/LIBPATH:C:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\libs', '/LIBPATH:C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.41.34120\\lib\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.22621.0\\ucrt\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.22621.0\\um\\x64', 'cuda.lib', '/OUT:C:\\Users\\Furkan\\AppData\\Local\\Temp\\tmpob5k0vz6\\cuda_utils.cp310-win_amd64.pyd']' returned non-zero exit status 2.

(venv) C:\Users\Furkan\Pictures\ghgfhfg>

subprocess.CalledProcessError: Command '['C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.41.34120\\bin\\Hostx64\\x64\\cl.EXE', 'C:\\Users\\Furkan\\AppData\\Local\\Temp\\tmpob5k0vz6\\main.c', '/nologo', '/O2', '/LD', '/wd4819', '/IC:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\Lib\\site-packages\\triton\\backends\\nvidia\\include', '/IC:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.4\\include', '/IC:\\Users\\Furkan\\AppData\\Local\\Temp\\tmpob5k0vz6', '/IC:\\Python310\\Include', '/IC:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.41.34120\\include', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\shared', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\ucrt', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\um', '/link', '/LIBPATH:C:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\Lib\\site-packages\\triton\\backends\\nvidia\\lib', '/LIBPATH:C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.4\\lib\\x64', '/LIBPATH:C:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\libs', '/LIBPATH:C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.41.34120\\lib\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.22621.0\\ucrt\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.22621.0\\um\\x64', 'cuda.lib', '/OUT:C:\\Users\\Furkan\\AppData\\Local\\Temp\\tmpob5k0vz6\\cuda_utils.cp310-win_amd64.pyd']' returned non-zero exit status 2.

@woct0rdho
Copy link
Owner Author

woct0rdho commented Oct 15, 2024

I see, you have Windows SDK in your "Visual Studio Build Tools", but not "Visual Studio Community". The easiest way is to also install Windows SDK in your "Visual Studio Community", and I suggest completely uninstall your "Visual Studio Build Tools"

@WingeD123
Copy link

got error:
File "D:\sd-ComfyUI\python_embeded\Lib\site-packages\triton\runtime\build.py", line 70, in _build
ret = subprocess.check_call(cc_cmd)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "subprocess.py", line 413, in check_call
torch._dynamo.exc.BackendCompilerFailed: backend='inductor' raised:
CalledProcessError: Command '['C:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.41.34120\bin\Hostx64\x64\cl.EXE', 'C:\Users\PLAY\AppData\Local\Temp\tmpvlvzc7z7\main.c', '/nologo', '/O2', '/LD', '/wd4819', '/ID:\sd-ComfyUI\python_embeded\Lib\site-packages\triton\backends\nvidia\include', '/IC:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.5\include', '/IC:\Users\PLAY\AppData\Local\Temp\tmpvlvzc7z7', '/ID:\sd-ComfyUI\python_embeded\Include', '/IC:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.41.34120\include', '/IC:\Program Files (x86)\Windows Kits\10\Include\10.0.20348.0\shared', '/IC:\Program Files (x86)\Windows Kits\10\Include\10.0.20348.0\ucrt', '/IC:\Program Files (x86)\Windows Kits\10\Include\10.0.20348.0\um', '/link', '/LIBPATH:D:\sd-ComfyUI\python_embeded\Lib\site-packages\triton\backends\nvidia\lib', '/LIBPATH:C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.5\lib\x64', '/LIBPATH:D:\sd-ComfyUI\python_embeded\libs', '/LIBPATH:C:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.41.34120\lib\x64', '/LIBPATH:C:\Program Files (x86)\Windows Kits\10\Lib\10.0.20348.0\ucrt\x64', '/LIBPATH:C:\Program Files (x86)\Windows Kits\10\Lib\10.0.20348.0\um\x64', '/LIBPATH:D:\sd-ComfyUI\python_embeded\libs', '/LIBPATH:C:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.41.34120\lib\x64', '/LIBPATH:C:\Program Files (x86)\Windows Kits\10\Lib\10.0.20348.0\ucrt\x64', '/LIBPATH:C:\Program Files (x86)\Windows Kits\10\Lib\10.0.20348.0\um\x64', 'cuda.lib', '/OUT:C:\Users\PLAY\AppData\Local\Temp\tmpvlvzc7z7\cuda_utils.cp311-win_amd64.pyd']' returned non-zero exit status 2.

@FurkanGozukara
Copy link

FurkanGozukara commented Oct 15, 2024

I see, you have Windows SDK in your "Visual Studio Build Tools", but not "Visual Studio Community". The easiest way is to also install Windows SDK in your "Visual Studio Community"

dmn i didnt show this on video tutorial. we can't fix it without install?

Visual Studio Build Tools is being required to compile so many stuff

installing SDK right now to test

@woct0rdho
Copy link
Owner Author

@woct0rdho
Copy link
Owner Author

@FurkanGozukara Modify C:\Python310\Lib\site-packages\triton\runtime\windows.py and try again

@@ -33,6 +33,8 @@
         "*",
         "-requires",
         "Microsoft.VisualStudio.Component.VC.Tools.x86.x64",
+        "-requires",
+        "Microsoft.VisualStudio.Component.Windows10SDK",
         "-latest",
         "-property",
         "installationPath",

@FurkanGozukara
Copy link

@woct0rdho installed SDK of win11 and made that change didnt fix

now installing windows 10 sdk too

@WingeD123
Copy link

@WingeD123 Please run the simple test script here: https://github.com/woct0rdho/triton?tab=readme-ov-file#test-if-it-works

subprocess.CalledProcessError: Command '['C:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.41.34120\bin\Hostx64\x64\cl.EXE', 'C:\Users\PLAY\AppData\Local\Temp\tmp525mv19u\main.c', '/nologo', '/O2', '/LD', '/wd4819', '/ID:\sd-ComfyUI\python_embeded\Lib\site-packages\triton\backends\nvidia\include', '/IC:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.5\include', '/IC:\Users\PLAY\AppData\Local\Temp\tmp525mv19u', '/ID:\sd-ComfyUI\python_embeded\Include', '/IC:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.41.34120\include', '/IC:\Program Files (x86)\Windows Kits\10\Include\10.0.20348.0\shared', '/IC:\Program Files (x86)\Windows Kits\10\Include\10.0.20348.0\ucrt', '/IC:\Program Files (x86)\Windows Kits\10\Include\10.0.20348.0\um', '/link', '/LIBPATH:D:\sd-ComfyUI\python_embeded\Lib\site-packages\triton\backends\nvidia\lib', '/LIBPATH:C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.5\lib\x64', '/LIBPATH:D:\sd-ComfyUI\python_embeded\libs', '/LIBPATH:C:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.41.34120\lib\x64', '/LIBPATH:C:\Program Files (x86)\Windows Kits\10\Lib\10.0.20348.0\ucrt\x64', '/LIBPATH:C:\Program Files (x86)\Windows Kits\10\Lib\10.0.20348.0\um\x64', 'cuda.lib', '/OUT:C:\Users\PLAY\AppData\Local\Temp\tmp525mv19u\cuda_utils.cp311-win_amd64.pyd']' returned non-zero exit status 2.

@FurkanGozukara
Copy link

FurkanGozukara commented Oct 15, 2024

@woct0rdho same error - i hate c++ tools - i only restarted the CMD not computer

image

image

(venv) C:\Users\Furkan\Pictures\ghgfhfg>python test.py
main.c
LINK : fatal error LNK1104: cannot open file 'python310.lib'
Traceback (most recent call last):
  File "C:\Users\Furkan\Pictures\ghgfhfg\test.py", line 26, in <module>
    b_compiled = add(a, a)
  File "C:\Users\Furkan\Pictures\ghgfhfg\test.py", line 21, in add
    add_kernel[grid](x, y, output, n_elements, BLOCK_SIZE=1024)
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\runtime\jit.py", line 345, in <lambda>
    return lambda *args, **kwargs: self.run(grid=grid, warmup=False, *args, **kwargs)
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\runtime\jit.py", line 607, in run
    device = driver.active.get_current_device()
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\runtime\driver.py", line 23, in __getattr__
    self._initialize_obj()
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\runtime\driver.py", line 20, in _initialize_obj
    self._obj = self._init_fn()
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\runtime\driver.py", line 9, in _create_driver
    return actives[0]()
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\backends\nvidia\driver.py", line 412, in __init__
    self.utils = CudaUtils()  # TODO: make static
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\backends\nvidia\driver.py", line 90, in __init__
    mod = compile_module_from_src(Path(os.path.join(dirname, "driver.c")).read_text(), "cuda_utils")
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\backends\nvidia\driver.py", line 67, in compile_module_from_src
    so = _build(name, src_path, tmpdir, library_dirs(), include_dir, libraries)
  File "C:\Users\Furkan\Pictures\ghgfhfg\venv\lib\site-packages\triton\runtime\build.py", line 75, in _build
    ret = subprocess.check_call(cc_cmd)
  File "C:\Python310\lib\subprocess.py", line 369, in check_call
    raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['C:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.38.33130\\bin\\Hostx64\\x64\\cl.EXE', 'C:\\Users\\Furkan\\AppData\\Local\\Temp\\tmp6xoqubwp\\main.c', '/nologo', '/O2', '/LD', '/wd4819', '/IC:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\Lib\\site-packages\\triton\\backends\\nvidia\\include', '/IC:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.4\\include', '/IC:\\Users\\Furkan\\AppData\\Local\\Temp\\tmp6xoqubwp', '/IC:\\Python310\\Include', '/IC:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.38.33130\\include', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.26100.0\\shared', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.26100.0\\ucrt', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.26100.0\\um', '/link', '/LIBPATH:C:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\Lib\\site-packages\\triton\\backends\\nvidia\\lib', '/LIBPATH:C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.4\\lib\\x64', '/LIBPATH:C:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\libs', '/LIBPATH:C:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.38.33130\\lib\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.26100.0\\ucrt\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.26100.0\\um\\x64', 'cuda.lib', '/OUT:C:\\Users\\Furkan\\AppData\\Local\\Temp\\tmp6xoqubwp\\cuda_utils.cp310-win_amd64.pyd']' returned non-zero exit status 2.

(venv) C:\Users\Furkan\Pictures\ghgfhfg>

@woct0rdho
Copy link
Owner Author

@FurkanGozukara Please run this in the same Python venv, and show the results:

import sysconfig
print(sysconfig.get_paths())

@FurkanGozukara
Copy link

FurkanGozukara commented Oct 15, 2024

@woct0rdho

by the way with my current system compiling InsightFace and XPose works

{'stdlib': 'C:\\Python310\\Lib', 'platstdlib': 'C:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\Lib', 'purelib': 'C:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\Lib\\site-packages', 'platlib': 'C:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\Lib\\site-packages', 'include': 'C:\\Python310\\Include', 'platinclude': 'C:\\Python310\\Include', 'scripts': 'C:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv\\Scripts', 'data': 'C:\\Users\\Furkan\\Pictures\\ghgfhfg\\venv'}

@FurkanGozukara
Copy link

let me test on c drive directly

@shivshankar11
Copy link

onediff>>> import sysconfig

print(sysconfig.get_paths())
{'stdlib': 'S:\Ai\Visual\StabilityMatrix\Data\Packages\ComfyUI 2\venv\Scripts\Lib', 'platstdlib': 'S:\Ai\Visual\StabilityMatrix\Data\Packages\ComfyUI 2\venv\Lib', 'purelib': 'S:\Ai\Visual\StabilityMatrix\Data\Packages\ComfyUI 2\venv\Lib\site-packages', 'platlib': 'S:\Ai\Visual\StabilityMatrix\Data\Packages\ComfyUI 2\venv\Lib\site-packages', 'include': 'S:\Ai\Visual\StabilityMatrix\Data\Packages\ComfyUI 2\venv\Scripts\Include', 'platinclude': 'S:\Ai\Visual\StabilityMatrix\Data\Packages\ComfyUI 2\venv\Scripts\Include', 'scripts': 'S:\Ai\Visual\StabilityMatrix\Data\Packages\ComfyUI 2\venv\Scripts', 'data': 'S:\Ai\Visual\StabilityMatrix\Data\Packages\ComfyUI 2\venv'}

@woct0rdho
Copy link
Owner Author

@FurkanGozukara Do you have these two folders: C:\Users\Furkan\Pictures\ghgfhfg\venv\libs, C:\Python310\libs? Please show all contents in them

Note that it's 'libs', not 'lib'

@FurkanGozukara
Copy link

FurkanGozukara commented Oct 15, 2024

@woct0rdho
no such folder as C:\Users\Furkan\Pictures\ghgfhfg\venv\libs

only lib is there

image

python folder has libs

image

@woct0rdho
Copy link
Owner Author

@FurkanGozukara Please modify C:\Python310\Lib\site-packages\triton\runtime\build.py and try again

@@ -63,6 +68,9 @@
     include_dirs = include_dirs + [srcdir, py_include_dir]
     if os.name == "nt":
         library_dirs += [os.path.join(sysconfig.get_paths()["data"], "libs")]
+        library_dirs += [os.path.join(os.path.dirname(sys.executable), "libs")]
+        python_version = sysconfig.get_python_version().replace(".", "")
+        library_dirs += [fr"C:\Python{python_version}\libs"]
         msvc_winsdk_inc_dirs, msvc_winsdk_lib_dirs = find_msvc_winsdk()
         include_dirs += msvc_winsdk_inc_dirs
         library_dirs += msvc_winsdk_lib_dirs

@FurkanGozukara
Copy link

FurkanGozukara commented Oct 15, 2024

@woct0rdho


(venv) C:\Users\Furkan\Pictures\ghgfhfg>python test.py
main.c
   Creating library main.lib and object main.exp
ptxas info    : 0 bytes gmem
ptxas info    : Compiling entry function 'add_kernel' for 'sm_86'
ptxas info    : Function properties for add_kernel
    0 bytes stack frame, 0 bytes spill stores, 0 bytes spill loads
ptxas info    : Used 26 registers, 380 bytes cmem[0]
main.c
   Creating library main.lib and object main.exp
tensor([0., 0., 0.], device='cuda:0')

@woct0rdho
Copy link
Owner Author

Great, the test script passed for you. You can try to do more things in ComfyUI

Now I'll make new wheels and let others try them

@FurkanGozukara
Copy link

Great, the test script passed for you. You can try to do more things in ComfyUI

Now I'll make new wheels and let others try them

yes but it is hard coded :D

i am waiting new wheel to test. i need this for general public users that follows me

@bghira
Copy link

bghira commented Oct 16, 2024

doesn't change anything about torch.compile, right? what is the point of having whl for windows ?

@woct0rdho
Copy link
Owner Author

woct0rdho commented Oct 16, 2024

@bghira Triton is the default backend of torch.compile. If Triton doesn't work, then torch.compile doesn't work by default, that's why we usually say "torch.compile doesn't work on Windows".

Actually it's possible to make Triton work on Windows, and what I'm doing here is to publish wheels and make it easier for more people to use it

@woct0rdho
Copy link
Owner Author

@NeoAnthropocene In your case you can successfully import triton (while some people see ImportError: DLL load failed when importing libtriton.pyd), and the problem happens when you load the newly compiled code

If you'd like you can help more with debugging. Install dlltracer in the embed Python environment, then in an administrator PowerShell, run the following:

import torch
import triton
import triton.language as tl

@triton.jit
def add_kernel(x_ptr, y_ptr, output_ptr, n_elements, BLOCK_SIZE: tl.constexpr):
    pid = tl.program_id(axis=0)
    block_start = pid * BLOCK_SIZE
    offsets = block_start + tl.arange(0, BLOCK_SIZE)
    mask = offsets < n_elements
    x = tl.load(x_ptr + offsets, mask=mask)
    y = tl.load(y_ptr + offsets, mask=mask)
    output = x + y
    tl.store(output_ptr + offsets, output, mask=mask)

def add(x: torch.Tensor, y: torch.Tensor):
    output = torch.empty_like(x)
    assert x.is_cuda and y.is_cuda and output.is_cuda
    n_elements = output.numel()
    grid = lambda meta: (triton.cdiv(n_elements, meta["BLOCK_SIZE"]),)
    add_kernel[grid](x, y, output, n_elements, BLOCK_SIZE=1024)
    return output

import sys
import dlltracer
with dlltracer.Trace(out=sys.stdout):
    a = torch.rand(3, device="cuda")
    b = a + a
    b_compiled = add(a, a)
    print(b_compiled - b)

@woct0rdho
Copy link
Owner Author

woct0rdho commented Oct 16, 2024

If you see Failed ... msvcp140.dll from dlltracer, that means you need to install vcredist (also known as 'Visual C++ Redistributable for Visual Studio 2015-2022', msvcp140.dll, vcruntime140.dll):
https://aka.ms/vs/17/release/vc_redist.x64.exe

The embeded Python already bundles vcruntime140.dll, but it's not the latest version. After installing the latest version, copy-paste msvcp140.dll, vcruntime140.dll, and vcruntime140_1.dll from C:\Windows\System32\ to the python_embeded folder

@NeoAnthropocene
Copy link

NeoAnthropocene commented Oct 16, 2024

@woct0rdho You're genius! Smashed the bug 🪲

Walkthrough to the debugging:

  1. Installed the dlltracer my ComfyUI virtual environment to catch the bug.
    F:\AI\ComfyUI_windows_portable\python_embeded\python.exe -m pip install dlltracer
  2. Run this on a Administrator PowerShell (Note that; I updated the test file as you gave here as tritonTest.py file inside embedded python folder of ComfyUI):
    F:\AI\ComfyUI_windows_portable\python_embeded\python.exe -m tritonTest
  3. I understand that I have a problem in that location failed. SO I deleted the old cached cuda_utils file under this location (actually I deleted the whole .triton folder)
PS C:\WINDOWS\system32> F:\AI\ComfyUI_windows_portable\python_embeded\python.exe -m tritonTest.py
LoadLibrary \Device\HarddiskVolume3\Windows\System32\kernel.appcore.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\DXCore.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\DriverStore\FileRepository\nv_dispi.inf_amd64_78cd02ab022cd554\nvcuda64.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\shlwapi.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\msasn1.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\cryptnet.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\wldp.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\drvstore.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\nvapi64.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\setupapi.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\DriverStore\FileRepository\nv_dispi.inf_amd64_78cd02ab022cd554\nvdxgdmal64.dll
LoadLibrary \Device\HarddiskVolume3\Users\altay\.triton\cache\f586be7c709ab2a9f1ad37d3d552072aa7ffbb2b0baaca2e9eca8d2b05ac1e22\cuda_utils.pyd
Failed \Device\HarddiskVolume3\Users\altay\.triton\cache\f586be7c709ab2a9f1ad37d3d552072aa7ffbb2b0baaca2e9eca8d2b05ac1e22\cuda_utils.pyd
  1. Then, I re-run the code again:
    PS C:\WINDOWS\system32> F:\AI\ComfyUI_windows_portable\python_embeded\python.exe -m tritonTest

That worked 😄

main.c
   Creating library main.lib and object main.exp
ptxas info    : 0 bytes gmem
ptxas info    : Compiling entry function 'add_kernel' for 'sm_89'
ptxas info    : Function properties for add_kernel
    0 bytes stack frame, 0 bytes spill stores, 0 bytes spill loads
ptxas info    : Used 26 registers, 380 bytes cmem[0]
main.c
   Creating library main.lib and object main.exp
tensor([0., 0., 0.], device='cuda:0')
LoadLibrary \Device\HarddiskVolume3\Windows\System32\kernel.appcore.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\DXCore.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\DriverStore\FileRepository\nv_dispi.inf_amd64_78cd02ab022cd554\nvcuda64.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\shlwapi.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\msasn1.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\cryptnet.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\wldp.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\drvstore.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\nvapi64.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\setupapi.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\DriverStore\FileRepository\nv_dispi.inf_amd64_78cd02ab022cd554\nvdxgdmal64.dll
LoadLibrary \Device\HarddiskVolume3\Users\altay\.triton\cache\f586be7c709ab2a9f1ad37d3d552072aa7ffbb2b0baaca2e9eca8d2b05ac1e22\cuda_utils.pyd
LoadLibrary \Device\HarddiskVolume3\Users\altay\.triton\cache\a799805a38a127e1610cefb76a44461750fb9779e70ee6031d95f36bcdb97d6e\__triton_launcher.pyd
  1. Finally, I turned the test script to your original one to see the money shot.
    🥳 Voila!
PS C:\WINDOWS\system32> F:\AI\ComfyUI_windows_portable\python_embeded\python.exe -m tritonTest
tensor([0., 0., 0.], device='cuda:0')
If you see tensor([0., 0., 0.], device='cuda:0'), then it works

@NeoAnthropocene In your case you can successfully import triton (while some people see ImportError: DLL load failed when importing libtriton.pyd), and the problem happens when you load the newly compiled code

If you'd like you can help more with debugging. Install dlltracer in the embed Python environment, then in an administrator PowerShell, run the following:

import torch
import triton
import triton.language as tl

@triton.jit
def add_kernel(x_ptr, y_ptr, output_ptr, n_elements, BLOCK_SIZE: tl.constexpr):
    pid = tl.program_id(axis=0)
    block_start = pid * BLOCK_SIZE
    offsets = block_start + tl.arange(0, BLOCK_SIZE)
    mask = offsets < n_elements
    x = tl.load(x_ptr + offsets, mask=mask)
    y = tl.load(y_ptr + offsets, mask=mask)
    output = x + y
    tl.store(output_ptr + offsets, output, mask=mask)

def add(x: torch.Tensor, y: torch.Tensor):
    output = torch.empty_like(x)
    assert x.is_cuda and y.is_cuda and output.is_cuda
    n_elements = output.numel()
    grid = lambda meta: (triton.cdiv(n_elements, meta["BLOCK_SIZE"]),)
    add_kernel[grid](x, y, output, n_elements, BLOCK_SIZE=1024)
    return output

import sys
import dlltracer
with dlltracer.Trace(out=sys.stdout):
    a = torch.rand(3, device="cuda")
    b = a + a
    b_compiled = add(a, a)
    print(b_compiled - b)

@Phosay
Copy link

Phosay commented Oct 16, 2024

I am getting quite curious error when trying to run anything to do with torch.compile using comfyUI, after getting VS and CUDA toolkit all set up, putting libs, include folders and installing triton wheels; backend inductor tries to access cache that already exists (FileExistsError: WinError 183) in the temp folder within AppData/Local.

@shivshankar11
Copy link

@woct0rdho You're genius! Smashed the bug 🪲

Walkthrough to the debugging:

  1. Installed the dlltracer my ComfyUI virtual environment to catch the bug.
    F:\AI\ComfyUI_windows_portable\python_embeded\python.exe -m pip install dlltracer
  2. Run this on a Administrator PowerShell (Note that; I updated the test file as you gave here as tritonTest.py file inside embedded python folder of ComfyUI):
    F:\AI\ComfyUI_windows_portable\python_embeded\python.exe -m tritonTest
  3. I understand that I have a problem in that location failed. SO I deleted the old cached cuda_utils file under this location (actually I deleted the whole .triton folder)
PS C:\WINDOWS\system32> F:\AI\ComfyUI_windows_portable\python_embeded\python.exe -m tritonTest.py
LoadLibrary \Device\HarddiskVolume3\Windows\System32\kernel.appcore.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\DXCore.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\DriverStore\FileRepository\nv_dispi.inf_amd64_78cd02ab022cd554\nvcuda64.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\shlwapi.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\msasn1.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\cryptnet.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\wldp.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\drvstore.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\nvapi64.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\setupapi.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\DriverStore\FileRepository\nv_dispi.inf_amd64_78cd02ab022cd554\nvdxgdmal64.dll
LoadLibrary \Device\HarddiskVolume3\Users\altay\.triton\cache\f586be7c709ab2a9f1ad37d3d552072aa7ffbb2b0baaca2e9eca8d2b05ac1e22\cuda_utils.pyd
Failed \Device\HarddiskVolume3\Users\altay\.triton\cache\f586be7c709ab2a9f1ad37d3d552072aa7ffbb2b0baaca2e9eca8d2b05ac1e22\cuda_utils.pyd
  1. Then, I re-run the code again:
    PS C:\WINDOWS\system32> F:\AI\ComfyUI_windows_portable\python_embeded\python.exe -m tritonTest

That worked 😄

main.c
   Creating library main.lib and object main.exp
ptxas info    : 0 bytes gmem
ptxas info    : Compiling entry function 'add_kernel' for 'sm_89'
ptxas info    : Function properties for add_kernel
    0 bytes stack frame, 0 bytes spill stores, 0 bytes spill loads
ptxas info    : Used 26 registers, 380 bytes cmem[0]
main.c
   Creating library main.lib and object main.exp
tensor([0., 0., 0.], device='cuda:0')
LoadLibrary \Device\HarddiskVolume3\Windows\System32\kernel.appcore.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\DXCore.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\DriverStore\FileRepository\nv_dispi.inf_amd64_78cd02ab022cd554\nvcuda64.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\shlwapi.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\msasn1.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\cryptnet.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\wldp.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\drvstore.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\nvapi64.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\setupapi.dll
LoadLibrary \Device\HarddiskVolume3\Windows\System32\DriverStore\FileRepository\nv_dispi.inf_amd64_78cd02ab022cd554\nvdxgdmal64.dll
LoadLibrary \Device\HarddiskVolume3\Users\altay\.triton\cache\f586be7c709ab2a9f1ad37d3d552072aa7ffbb2b0baaca2e9eca8d2b05ac1e22\cuda_utils.pyd
LoadLibrary \Device\HarddiskVolume3\Users\altay\.triton\cache\a799805a38a127e1610cefb76a44461750fb9779e70ee6031d95f36bcdb97d6e\__triton_launcher.pyd
  1. Finally, I turned the test script to your original one to see the money shot.
    🥳 Voila!
PS C:\WINDOWS\system32> F:\AI\ComfyUI_windows_portable\python_embeded\python.exe -m tritonTest
tensor([0., 0., 0.], device='cuda:0')
If you see tensor([0., 0., 0.], device='cuda:0'), then it works

@NeoAnthropocene In your case you can successfully import triton (while some people see ImportError: DLL load failed when importing libtriton.pyd), and the problem happens when you load the newly compiled code
If you'd like you can help more with debugging. Install dlltracer in the embed Python environment, then in an administrator PowerShell, run the following:

import torch
import triton
import triton.language as tl

@triton.jit
def add_kernel(x_ptr, y_ptr, output_ptr, n_elements, BLOCK_SIZE: tl.constexpr):
    pid = tl.program_id(axis=0)
    block_start = pid * BLOCK_SIZE
    offsets = block_start + tl.arange(0, BLOCK_SIZE)
    mask = offsets < n_elements
    x = tl.load(x_ptr + offsets, mask=mask)
    y = tl.load(y_ptr + offsets, mask=mask)
    output = x + y
    tl.store(output_ptr + offsets, output, mask=mask)

def add(x: torch.Tensor, y: torch.Tensor):
    output = torch.empty_like(x)
    assert x.is_cuda and y.is_cuda and output.is_cuda
    n_elements = output.numel()
    grid = lambda meta: (triton.cdiv(n_elements, meta["BLOCK_SIZE"]),)
    add_kernel[grid](x, y, output, n_elements, BLOCK_SIZE=1024)
    return output

import sys
import dlltracer
with dlltracer.Trace(out=sys.stdout):
    a = torch.rand(3, device="cuda")
    b = a + a
    b_compiled = add(a, a)
    print(b_compiled - b)

thanks, worked

@woct0rdho
Copy link
Owner Author

@Phosay Please paste the whole error log, not only the last line

@jepjoo
Copy link

jepjoo commented Oct 16, 2024

I am getting quite curious error when trying to run anything to do with torch.compile using comfyUI, after getting VS and CUDA toolkit all set up, putting libs, include folders and installing triton wheels; backend inductor tries to access cache that already exists (FileExistsError: WinError 183) in the temp folder within AppData/Local.

This sounds like the error I'm getting when using torch.compile with Flux and CFG higher than 1. CFG = 1.0 works fine.

I think with CFG 1< it tries to do the torch.compile twice and because the file already exists for the first one, it fails on the second one.

On WSL, CFG 1< required the second compile aswell, but it did still work.

I don't have the error for this saved, can get back to this tomorrow if @Phosay doesn't do it first (assuming this is the same issue).

@frankyifei
Copy link

From my test, it give me 10% speed increase on my 3060

@Telllinex
Copy link

From my test, it give me 10% speed increase on my 3060

What tasks? ComfyUI or Forge? i ran test.py and it works directly from Forge venv, but in generating images there is no difference

@woct0rdho
Copy link
Owner Author

I am getting quite curious error when trying to run anything to do with torch.compile using comfyUI, after getting VS and CUDA toolkit all set up, putting libs, include folders and installing triton wheels; backend inductor tries to access cache that already exists (FileExistsError: WinError 183) in the temp folder within AppData/Local.

@Phosay You need to modify the file torch/_inductor/codecache.py, around line 404, in function write_atomic, change tmp_path.remove to tmp_path.replace

See pytorch/pytorch#138211

@jepjoo
Copy link

jepjoo commented Oct 18, 2024

Flux CFG at 1.0 works fine, Flux CFG higher than 1, I get this:

got prompt 0%| | 0/30 [00:14 'C:\\Users\\USERNAME\\AppData\\Local\\Temp\\torchinductor_USERNAME\\cache\\84b8dc1bae2f40b2751b45cbdfa5721ab7a12992b441e7590a3e9f612e421f8a'

Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information

You can suppress this exception and fall back to eager by setting:
import torch._dynamo
torch._dynamo.config.suppress_errors = True

Traceback (most recent call last):
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\output_graph.py", line 1446, in call_user_compiler
compiled_fn = compiler_fn(gm, self.example_inputs())
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\repro\after_dynamo.py", line 129, in call
compiled_gm = compiler_fn(gm, example_inputs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_init
.py", line 2235, in call
return compile_fx(model_, inputs_, config_patches=self.config)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\compile_fx.py", line 1521, in compile_fx
return aot_autograd(
^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\backends\common.py", line 72, in call
cg = aot_module_simplified(gm, example_inputs, **self.kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_functorch\aot_autograd.py", line 1071, in aot_module_simplified
compiled_fn = dispatch_and_compile()
^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_functorch\aot_autograd.py", line 1056, in dispatch_and_compile
compiled_fn, _ = create_aot_dispatcher_function(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_functorch\aot_autograd.py", line 522, in create_aot_dispatcher_function
return _create_aot_dispatcher_function(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_functorch\aot_autograd.py", line 759, in _create_aot_dispatcher_function
compiled_fn, fw_metadata = compiler_fn(
^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_functorch_aot_autograd\jit_compile_runtime_wrappers.py", line 179, in aot_dispatch_base
compiled_fw = compiler(fw_module, updated_flat_args)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\compile_fx.py", line 1350, in fw_compiler_base
return _fw_compiler_base(model, example_inputs, is_inference)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\compile_fx.py", line 1359, in _fw_compiler_base
_recursive_joint_graph_passes(model)
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\compile_fx.py", line 281, in _recursive_joint_graph_passes
joint_graph_passes(gm)
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\fx_passes\joint_graph.py", line 460, in joint_graph_passes
count += patterns.apply(graph.graph) # type: ignore[arg-type]
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\pattern_matcher.py", line 1729, in apply
if is_match(m) and entry.extra_check(m):
^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\pattern_matcher.py", line 1331, in check_fn
if is_match(specific_pattern_match) and extra_check(specific_pattern_match):
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\fx_passes\pad_mm.py", line 146, in should_pad_addmm
return should_pad_common(mat1, mat2, input) and should_pad_bench(
^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\fx_passes\pad_mm.py", line 567, in should_pad_bench
set_cached_base_mm_benchmark_time(ori_time_key, ori_time)
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\fx_passes\pad_mm.py", line 262, in set_cached_base_mm_benchmark_time
return get_pad_cache().set_value(key, value=value)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\codecache.py", line 286, in set_value
self.update_local_cache(cache)
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\codecache.py", line 257, in update_local_cache
write_atomic(
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_inductor\codecache.py", line 466, in write_atomic
tmp_path.rename(path)
File "pathlib.py", line 1175, in rename
FileExistsError: [WinError 183] Cannot create a file when that file already exists: 'C:\Users\USERNAME\AppData\Local\Temp\torchinductor_USERNAME\cache\.7416.7720.tmp' -> 'C:\Users\USERNAME\AppData\Local\Temp\torchinductor_USERNAME\cache\84b8dc1bae2f40b2751b45cbdfa5721ab7a12992b441e7590a3e9f612e421f8a'

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
File "G:\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in map_node_over_list
process_inputs(input_dict, i)
File "G:\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
results.append(getattr(obj, func)(**inputs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1471, in sample
return common_ksampler(model, noise_seed, steps, cfg, sampler_name, scheduler, positive, negative, latent_image, denoise=denoise, disable_noise=disable_noise, start_step=start_at_step, last_step=end_at_step, force_full_denoise=force_full_denoise)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1404, in common_ksampler
samples = comfy.sample.sample(model, noise, steps, cfg, sampler_name, scheduler, positive, negative, latent_image,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\sample_error_enhancer.py", line 22, in informative_sample
raise e
File "G:\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\sample_error_enhancer.py", line 9, in informative_sample
return original_sample(*args, **kwargs) # This code helps interpret error messages that occur within exceptions but does not have any impact on other operations.
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\sample.py", line 43, in sample
samples = sampler.sample(noise, positive, negative, cfg=cfg, latent_image=latent_image, start_step=start_step, last_step=last_step, force_full_denoise=force_full_denoise, denoise_mask=noise_mask, sigmas=sigmas, callback=callback, disable_pbar=disable_pbar, seed=seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 829, in sample
return sample(self.model, noise, positive, negative, cfg, self.device, sampler, sigmas, self.model_options, latent_image=latent_image, denoise_mask=denoise_mask, callback=callback, disable_pbar=disable_pbar, seed=seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 729, in sample
return cfg_guider.sample(noise, latent_image, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 716, in sample
output = self.inner_sample(noise, latent_image, device, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 695, in inner_sample
samples = sampler.sample(self, sigmas, extra_args, callback, noise, latent_image, denoise_mask, disable_pbar)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 600, in sample
samples = self.sampler_function(model_k, noise, sigmas, extra_args=extra_args, callback=k_callback, disable=disable_pbar, **self.extra_options)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\utils_contextlib.py", line 116, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\k_diffusion\sampling.py", line 155, in sample_euler
denoised = model(x, sigma_hat * s_in, **extra_args)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 299, in call
out = self.inner_model(x, sigma, model_options=model_options, seed=seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 682, in call
return self.predict_noise(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 685, in predict_noise
return sampling_function(self.inner_model, x, timestep, self.conds.get("negative", None), self.conds.get("positive", None), self.cfg, model_options=model_options, seed=seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 279, in sampling_function
out = calc_cond_batch(model, conds, x, timestep, model_options)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 228, in calc_cond_batch
output = model.apply_model(input_x, timestep
, **c).chunk(batch_chunks)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\ComfyUI\comfy\model_base.py", line 142, in apply_model
model_output = self.diffusion_model(xc, t, context=context, control=control, transformer_options=transformer_options, **extra_conds).float()
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\eval_frame.py", line 465, in _fn
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\convert_frame.py", line 1269, in call
return self._torchdynamo_orig_callable(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\convert_frame.py", line 1064, in call
result = self._inner_convert(
^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\convert_frame.py", line 526, in call
return _compile(
^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\convert_frame.py", line 924, in _compile
guarded_code = compile_inner(code, one_graph, hooks, transform)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\convert_frame.py", line 666, in compile_inner
return _compile_inner(code, one_graph, hooks, transform)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_utils_internal.py", line 87, in wrapper_function
return function(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\convert_frame.py", line 699, in _compile_inner
out_code = transform_code_object(code, transform)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\bytecode_transformation.py", line 1322, in transform_code_object
transformations(instructions, code_options)
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\convert_frame.py", line 219, in _fn
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\convert_frame.py", line 634, in transform
tracer.run()
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\symbolic_convert.py", line 2796, in run
super().run()
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\symbolic_convert.py", line 983, in run
while self.step():
^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\symbolic_convert.py", line 895, in step
self.dispatch_table[inst.opcode](self, inst)
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\symbolic_convert.py", line 2987, in RETURN_VALUE
self._return(inst)
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\symbolic_convert.py", line 2972, in _return
self.output.compile_subgraph(
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\output_graph.py", line 1117, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\output_graph.py", line 1369, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\output_graph.py", line 1416, in call_user_compiler
return self._call_user_compiler(gm)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch_dynamo\output_graph.py", line 1465, in _call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: backend='inductor' raised:
FileExistsError: [WinError 183] Cannot create a file when that file already exists: 'C:\Users\USERNAME\AppData\Local\Temp\torchinductor_USERNAME\cache\.7416.7720.tmp' -> 'C:\Users\USERNAME\AppData\Local\Temp\torchinductor_USERNAME\cache\84b8dc1bae2f40b2751b45cbdfa5721ab7a12992b441e7590a3e9f612e421f8a'

Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information

You can suppress this exception and fall back to eager by setting:
import torch._dynamo
torch._dynamo.config.suppress_errors = True

Prompt executed in 14.89 seconds

@woct0rdho
Copy link
Owner Author

@jepjoo You also need to modify this file, see pytorch/pytorch#138211

@jepjoo
Copy link

jepjoo commented Oct 18, 2024

That fixed it thanks!

@shivshankar11
Copy link

Flux CFG at 1.0 works fine, Flux CFG higher than 1, I get this:

ca you use batch size 2 with torch.compile? i get same error if batch size is not 1.

@jepjoo
Copy link

jepjoo commented Oct 18, 2024

Flux CFG at 1.0 works fine, Flux CFG higher than 1, I get this:

ca you use batch size 2 with torch.compile? i get same error if batch size is not 1.

Just tried it and yes, batch size 2 works with torch.compile for me.

Try the fix above. It's quite simple.

@shivshankar11
Copy link

shivshankar11 commented Oct 18, 2024

Flux CFG at 1.0 works fine, Flux CFG higher than 1, I get this:

ca you use batch size 2 with torch.compile? i get same error if batch size is not 1.

Just tried it and yes, batch size 2 works with torch.compile for me.

Try the fix above. It's quite simple.

i am using 4090, after fix batch 2 is crashing comfyui

@NikosKont
Copy link

Great work, any chance for a python3.9 wheel?

@woct0rdho
Copy link
Owner Author

@NikosKont Sure, now the wheels for Python 3.8 and 3.9 are published. I did not fully test them, but they should work because the official Triton publishes wheels for Python 3.8 to 3.12

@maxisoft
Copy link

I tested these Triton wheels to run PyTorch FlexAttention on Windows 10. It seems to work, thanks for your work!

Sharing a note: while trying to compile a TanhApprox function, I encountered a missing CUDA DLL due to a mismatch between PyTorch compiled with cu12.4 and my installed CUDA version 12.6. To resolve this, I created a symbolic link nvrtc-builtins64_124.dll -> nvrtc-builtins64_126.dll.
while the root of the issue is a cuda version missmatch and internal pytorch FlexAttention compilation directives, one may found the solution useful.

@mityarko
Copy link

I am getting quite curious error when trying to run anything to do with torch.compile using comfyUI, after getting VS and CUDA toolkit all set up, putting libs, include folders and installing triton wheels; backend inductor tries to access cache that already exists (FileExistsError: WinError 183) in the temp folder within AppData/Local.

@Phosay You need to modify the file torch/_inductor/codecache.py, around line 404, in function write_atomic, change tmp_path.remove to tmp_path.replace

See pytorch/pytorch#138211

Also faced with this, will hope for a wheel package update soon

@Ratinod
Copy link

Ratinod commented Nov 2, 2024

Has anyone encountered the problem of colored cubes in CogVideoX?
kijai/ComfyUI-CogVideoXWrapper#200

@cmp-nct
Copy link

cmp-nct commented Nov 2, 2024

image
On a 4090 with torch 2.5.0+cu124

Experimenting a bit more: This error happens when using a GGUF CLIP (Q6K) loader of the T5 XXL. I get a proper image when using a fp8 torch version of the same encoder.

@Ratinod
Copy link

Ratinod commented Nov 2, 2024

Experimenting a bit more: This error happens when using a GGUF CLIP (Q6K) loader of the T5 XXL. I get a proper image when using a fp8 torch version of the same encoder.

I shake your hand.
I spent the whole day trying to find the problem but I didn't even think that the problem could be in "ClipLoader (GGUF)" with the model "t5-v1_1-xxl-encoder-Q8_0.gguf"
But if we use "Load CLIP" and the model "t5xxl_fp8_e4m3fn.safetensors" then everything works.
The only question is why gguf clip doesn't work well and can it be fixed? (And on whose side is the fix required?)

@Danteday
Copy link

Danteday commented Nov 6, 2024

Hi everyone, I don't understand what the error is. I tried to edit the file torch/_inductor/codecache.py but there are no lines to replace.

Нажмите, чтобы раскрыть текст # ComfyUI Error Report ## Error Details - **Node Type:** KSampler - **Exception Type:** torch._dynamo.exc.BackendCompilerFailed - **Exception Message:** backend='inductor' raised: CompilationError: at 16:15: xindex = xoffset + tl.arange(0, XBLOCK)[:, None] xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1) rbase = tl.arange(0, RBLOCK)[None, :] x0 = xindex _tmp11 = tl.full([XBLOCK, RBLOCK], 0, tl.float32) _tmp18 = tl.full([XBLOCK, RBLOCK], 0, tl.float32) for roffset in range(0, rnumel, RBLOCK): rindex = roffset + rbase rmask = rindex < rnumel r1 = rindex tmp0 = tl.load(in_ptr0 + (r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) tmp6 = tl.load(in_ptr1 + (r1 + (3072*x0)), rmask, eviction_policy='evict_first', other=0.0) ^

Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information

You can suppress this exception and fall back to eager by setting:
import torch._dynamo
torch._dynamo.config.suppress_errors = True

Stack Trace

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
    process_inputs(input_dict, i)

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1442, in sample
    return common_ksampler(model, seed, steps, cfg, sampler_name, scheduler, positive, negative, latent_image, denoise=denoise)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1409, in common_ksampler
    samples = comfy.sample.sample(model, noise, steps, cfg, sampler_name, scheduler, positive, negative, latent_image,
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\sample_error_enhancer.py", line 22, in informative_sample
    raise e

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\sample_error_enhancer.py", line 9, in informative_sample
    return original_sample(*args, **kwargs)  # This code helps interpret error messages that occur within exceptions but does not have any impact on other operations.
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\sample.py", line 43, in sample
    samples = sampler.sample(noise, positive, negative, cfg=cfg, latent_image=latent_image, start_step=start_step, last_step=last_step, force_full_denoise=force_full_denoise, denoise_mask=noise_mask, sigmas=sigmas, callback=callback, disable_pbar=disable_pbar, seed=seed)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 855, in sample
    return sample(self.model, noise, positive, negative, cfg, self.device, sampler, sigmas, self.model_options, latent_image=latent_image, denoise_mask=denoise_mask, callback=callback, disable_pbar=disable_pbar, seed=seed)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 753, in sample
    return cfg_guider.sample(noise, latent_image, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 740, in sample
    output = self.inner_sample(noise, latent_image, device, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 719, in inner_sample
    samples = sampler.sample(self, sigmas, extra_args, callback, noise, latent_image, denoise_mask, disable_pbar)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 624, in sample
    samples = self.sampler_function(model_k, noise, sigmas, extra_args=extra_args, callback=k_callback, disable=disable_pbar, **self.extra_options)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\k_diffusion\sampling.py", line 155, in sample_euler
    denoised = model(x, sigma_hat * s_in, **extra_args)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 299, in __call__
    out = self.inner_model(x, sigma, model_options=model_options, seed=seed)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 706, in __call__
    return self.predict_noise(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 709, in predict_noise
    return sampling_function(self.inner_model, x, timestep, self.conds.get("negative", None), self.conds.get("positive", None), self.cfg, model_options=model_options, seed=seed)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 279, in sampling_function
    out = calc_cond_batch(model, conds, x, timestep, model_options)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 228, in calc_cond_batch
    output = model.apply_model(input_x, timestep_, **c).chunk(batch_chunks)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\model_base.py", line 144, in apply_model
    model_output = self.diffusion_model(xc, t, context=context, control=control, transformer_options=transformer_options, **extra_conds).float()
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\ldm\flux\model.py", line 159, in forward
    out = self.forward_orig(img, img_ids, context, txt_ids, timestep, y, guidance, control)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\ldm\flux\model.py", line 118, in forward_orig
    img, txt = block(img=img, txt=txt, vec=vec, pe=pe)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\eval_frame.py", line 465, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 1269, in __call__
    return self._torchdynamo_orig_callable(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 1064, in __call__
    result = self._inner_convert(
             ^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 526, in __call__
    return _compile(
           ^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 924, in _compile
    guarded_code = compile_inner(code, one_graph, hooks, transform)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 666, in compile_inner
    return _compile_inner(code, one_graph, hooks, transform)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_utils_internal.py", line 87, in wrapper_function
    return function(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 699, in _compile_inner
    out_code = transform_code_object(code, transform)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\bytecode_transformation.py", line 1322, in transform_code_object
    transformations(instructions, code_options)

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 219, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 634, in transform
    tracer.run()

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 2796, in run
    super().run()

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 983, in run
    while self.step():
          ^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 895, in step
    self.dispatch_table[inst.opcode](self, inst)

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 2987, in RETURN_VALUE
    self._return(inst)

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 2972, in _return
    self.output.compile_subgraph(

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1142, in compile_subgraph
    self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1369, in compile_and_call_fx_graph
    compiled_fn = self.call_user_compiler(gm)
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1416, in call_user_compiler
    return self._call_user_compiler(gm)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1465, in _call_user_compiler
    raise BackendCompilerFailed(self.compiler_fn, e) from e

System Information

  • ComfyUI Version: v0.2.7-3-g8afb97c
  • Arguments: ComfyUI\main.py --windows-standalone-build --fast
  • OS: nt
  • Python Version: 3.11.9 (tags/v3.11.9:de54cf5, Apr 2 2024, 10:12:12) [MSC v.1938 64 bit (AMD64)]
  • Embedded Python: true
  • PyTorch Version: 2.5.1+cu124

Devices

  • Name: cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync
    • Type: cuda
    • VRAM Total: 25769279488
    • VRAM Free: 12349091842
    • Torch VRAM Total: 12079595520
    • Torch VRAM Free: 66072578

Logs

2024-11-05 23:40:13,475 - root - INFO - Total VRAM 24576 MB, total RAM 65372 MB
2024-11-05 23:40:13,475 - root - INFO - pytorch version: 2.5.1+cu124
2024-11-05 23:40:13,476 - root - INFO - Set vram state to: NORMAL_VRAM
2024-11-05 23:40:13,476 - root - INFO - Device: cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync
2024-11-05 23:40:14,680 - root - INFO - Using pytorch cross attention
2024-11-05 23:40:15,518 - root - INFO - [Prompt Server] web root: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\web
2024-11-05 23:40:17,729 - root - INFO - Total VRAM 24576 MB, total RAM 65372 MB
2024-11-05 23:40:17,729 - root - INFO - pytorch version: 2.5.1+cu124
2024-11-05 23:40:17,730 - root - INFO - Set vram state to: NORMAL_VRAM
2024-11-05 23:40:17,730 - root - INFO - Device: cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync
2024-11-05 23:40:18,719 - root - WARNING - Traceback (most recent call last):
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\nodes.py", line 2012, in load_custom_node
    module_spec.loader.exec_module(module)
  File "<frozen importlib._bootstrap_external>", line 940, in exec_module
  File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Inspire-Pack\__init__.py", line 34, in <module>
    imported_module = importlib.import_module(".inspire.{}".format(module_name), __name__)
                      ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "importlib\__init__.py", line 126, in import_module
  File "<frozen importlib._bootstrap>", line 1204, in _gcd_import
  File "<frozen importlib._bootstrap>", line 1176, in _find_and_load
  File "<frozen importlib._bootstrap>", line 1147, in _find_and_load_unlocked
  File "<frozen importlib._bootstrap>", line 690, in _load_unlocked
  File "<frozen importlib._bootstrap_external>", line 940, in exec_module
  File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Inspire-Pack\inspire\regional_nodes.py", line 7, in <module>
    import webcolors
ModuleNotFoundError: No module named 'webcolors'

2024-11-05 23:40:18,720 - root - WARNING - Cannot import C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Inspire-Pack module for custom nodes: No module named 'webcolors'
2024-11-05 23:40:20,138 - root - INFO - --------------
2024-11-05 23:40:20,139 - root - INFO - �[91m ### Mixlab Nodes: �[93mLoaded
2024-11-05 23:40:20,147 - root - INFO - ChatGPT.available True
2024-11-05 23:40:20,148 - root - INFO - edit_mask.available True
2024-11-05 23:40:20,160 - root - INFO - ClipInterrogator.available True
2024-11-05 23:40:20,274 - root - INFO - PromptGenerate.available True
2024-11-05 23:40:20,274 - root - INFO - ChinesePrompt.available True
2024-11-05 23:40:20,274 - root - INFO - RembgNode_.available True
2024-11-05 23:40:20,600 - root - INFO - TripoSR.available
2024-11-05 23:40:20,601 - root - INFO - MiniCPMNode.available
2024-11-05 23:40:20,652 - root - INFO - Scenedetect.available
2024-11-05 23:40:20,724 - root - INFO - FishSpeech.available
2024-11-05 23:40:20,731 - root - INFO - SenseVoice.available
2024-11-05 23:40:20,750 - root - INFO - Whisper.available False
2024-11-05 23:40:20,758 - root - INFO - FalVideo.available
2024-11-05 23:40:20,758 - root - INFO - �[93m -------------- �[0m
2024-11-05 23:40:21,525 - py.warnings - WARNING - C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\albumentations\__init__.py:13: UserWarning: A new version of Albumentations is available: 1.4.21 (you have 1.4.18). Upgrade using: pip install -U albumentations. To disable automatic update checks, set the environment variable NO_ALBUMENTATIONS_UPDATE to 1.
  check_for_updates()

2024-11-05 23:40:21,949 - root - WARNING - Traceback (most recent call last):
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\nodes.py", line 2012, in load_custom_node
    module_spec.loader.exec_module(module)
  File "<frozen importlib._bootstrap_external>", line 940, in exec_module
  File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_alkaid\__init__.py", line 1, in <module>
    from .tool import Face3DSwapper, FaceCrop, FacePaste, ApplyAdapter, CombineAdapterPatch, KSamplerHires
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_alkaid\tool.py", line 11, in <module>
    from comfyui_controlnet_aux.utils import common_annotator_call, create_node_input_types
ImportError: cannot import name 'create_node_input_types' from 'comfyui_controlnet_aux.utils' (C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui_controlnet_aux\utils.py)

2024-11-05 23:40:21,949 - root - WARNING - Cannot import C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_alkaid module for custom nodes: cannot import name 'create_node_input_types' from 'comfyui_controlnet_aux.utils' (C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui_controlnet_aux\utils.py)
2024-11-05 23:40:23,373 - root - WARNING - Traceback (most recent call last):
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\nodes.py", line 2012, in load_custom_node
    module_spec.loader.exec_module(module)
  File "<frozen importlib._bootstrap_external>", line 940, in exec_module
  File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_HF_Servelress_Inference\__init__.py", line 9, in <module>
    from .nodes.Joy_Caption import JoyPipeline, Joy_caption, Joy_caption_load
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_HF_Servelress_Inference\nodes\Joy_Caption.py", line 171
    with torch.amp.autocast_mode.autocast(DEVICE.type if isinstance(DEVICE, torch.device) else DEVICE, enabled=True)
                                                                                                                    ^
SyntaxError: expected ':'

2024-11-05 23:40:23,373 - root - WARNING - Cannot import C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_HF_Servelress_Inference module for custom nodes: expected ':' (Joy_Caption.py, line 171)
2024-11-05 23:40:29,909 - root - INFO - 
Import times for custom nodes:
2024-11-05 23:40:29,910 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\lying_sigma_sampler.py
2024-11-05 23:40:29,910 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\websocket_image_save.py
2024-11-05 23:40:29,910 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-CustomScheduler
2024-11-05 23:40:29,910 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyUI_TJ_NormalLighting
2024-11-05 23:40:29,910 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\sdxl_prompt_styler
2024-11-05 23:40:29,910 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui-image-round
2024-11-05 23:40:29,910 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-SDXL-EmptyLatentImage
2024-11-05 23:40:29,910 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_SUNoise
2024-11-05 23:40:29,910 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\cg-use-everywhere
2024-11-05 23:40:29,910 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\efficiency-nodes-comfyui
2024-11-05 23:40:29,910 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-yaResolutionSelector
2024-11-05 23:40:29,910 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-mxToolkit
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Detail-Daemon
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Inpaint-CropAndStitch
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\lora-info
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\stability-ComfyUI-nodes
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_TiledKSampler
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-YOLO
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Eagle-PNGInfo
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\cg-image-picker
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-VideoHelperSuite
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Logic
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\masquerade-nodes-comfyui
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\sigmas_tools_and_the_golden_scheduler
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds (IMPORT FAILED): C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_alkaid
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyLiterals
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui_controlnet_aux
2024-11-05 23:40:29,911 - root - INFO -    0.0 seconds (IMPORT FAILED): C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_HF_Servelress_Inference
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_JPS-Nodes
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\Comfyui_CXH_joy_caption
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Florence2
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui-tooling-nodes
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Fluxtapoz
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Extra-Samplers
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfy-image-saver
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\DJZ-Nodes
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\mikey_nodes
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-AutoCropFaces
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui-various
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Custom-Scripts
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Florence-2
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_UltimateSDUpscale
2024-11-05 23:40:29,912 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\Derfuu_ComfyUI_ModdedNodes
2024-11-05 23:40:29,913 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\x-flux-comfyui
2024-11-05 23:40:29,913 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-KJNodes
2024-11-05 23:40:29,913 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_essentials
2024-11-05 23:40:29,913 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-GGUF
2024-11-05 23:40:29,913 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\rgthree-comfy
2024-11-05 23:40:29,913 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_ExtraModels
2024-11-05 23:40:29,913 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui-workspace-manager
2024-11-05 23:40:29,913 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_Comfyroll_CustomNodes
2024-11-05 23:40:29,913 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_Searge_LLM
2024-11-05 23:40:29,913 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\Comfyui-ergouzi-Nodes
2024-11-05 23:40:29,913 - root - INFO -    0.0 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-RvTools
2024-11-05 23:40:29,913 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_tinyterraNodes
2024-11-05 23:40:29,913 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools
2024-11-05 23:40:29,913 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Geowizard
2024-11-05 23:40:29,913 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_bnb_nf4_fp4_Loaders
2024-11-05 23:40:29,913 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Long-CLIP
2024-11-05 23:40:29,913 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui_segment_anything
2024-11-05 23:40:29,914 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui-reactor-node
2024-11-05 23:40:29,914 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-PuLID-Flux-Enhanced
2024-11-05 23:40:29,914 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui-tensorops
2024-11-05 23:40:29,914 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_LayerStyle
2024-11-05 23:40:29,914 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Easy-Use
2024-11-05 23:40:29,914 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-AdvancedLivePortrait
2024-11-05 23:40:29,914 - root - INFO -    0.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\DZ-FaceDetailer
2024-11-05 23:40:29,914 - root - INFO -    0.2 seconds (IMPORT FAILED): C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Inspire-Pack
2024-11-05 23:40:29,914 - root - INFO -    0.3 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-DepthCrafter-Nodes
2024-11-05 23:40:29,914 - root - INFO -    0.4 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui-art-venture
2024-11-05 23:40:29,914 - root - INFO -    0.4 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager
2024-11-05 23:40:29,914 - root - INFO -    0.5 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-LLaVA-OneVision
2024-11-05 23:40:29,914 - root - INFO -    0.7 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\bilbox-comfyui
2024-11-05 23:40:29,914 - root - INFO -    0.8 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Image-Filters
2024-11-05 23:40:29,914 - root - INFO -    0.9 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-PuLID-Flux
2024-11-05 23:40:29,914 - root - INFO -    1.1 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui-mixlab-nodes
2024-11-05 23:40:29,915 - root - INFO -    1.3 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\was-node-suite-comfyui
2024-11-05 23:40:29,915 - root - INFO -    1.3 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_Custom_Nodes_AlekPet
2024-11-05 23:40:29,915 - root - INFO -    4.8 seconds: C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui_LLM_party
2024-11-05 23:40:29,915 - root - INFO - 
2024-11-05 23:40:29,928 - root - INFO - Starting server

2024-11-05 23:40:29,928 - root - INFO - To see the GUI go to: http://127.0.0.1:8188
2024-11-05 23:41:26,810 - root - INFO - got prompt
2024-11-05 23:41:27,255 - root - WARNING - clip missing: ['text_projection.weight']
2024-11-05 23:41:28,522 - root - INFO - Requested to load FluxClipModel_
2024-11-05 23:41:28,522 - root - INFO - Loading 1 new model
2024-11-05 23:41:29,639 - root - INFO - loaded completely 0.0 4777.53759765625 True
2024-11-05 23:41:30,168 - root - INFO - Using pytorch attention in VAE
2024-11-05 23:41:30,169 - root - INFO - Using pytorch attention in VAE
2024-11-05 23:41:30,355 - root - INFO - model weight dtype torch.float8_e4m3fn, manual cast: torch.bfloat16
2024-11-05 23:41:30,356 - root - INFO - model_type FLUX
2024-11-05 23:41:35,494 - root - WARNING - Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
2024-11-05 23:41:35,510 - root - INFO - Requested to load Flux
2024-11-05 23:41:35,511 - root - INFO - Loading 1 new model
2024-11-05 23:41:39,970 - root - INFO - loaded completely 0.0 11350.048889160156 True
2024-11-05 23:42:21,324 - root - ERROR - !!! Exception during processing !!! backend='inductor' raised:
CompilationError: at 18:11:
    xindex = tl.full([1], xoffset, tl.int32)
    xmask = tl.full([RBLOCK], True, tl.int1)
    rindex = tl.arange(0, RBLOCK)[:]
    roffset = 0
    rmask = tl.full([RBLOCK], True, tl.int1)
    r1 = rindex
    x0 = xindex
    tmp0 = tl.load(in_ptr0 + (r1), None, eviction_policy='evict_last')
    tmp3 = tl.load(in_ptr1 + (r1 + (256*x0)), None)
    tmp1 = tmp0.to(tl.float32)
    tmp2 = tmp1.to(tl.float32)
    tmp4 = tmp3.to(tl.float32)
           ^

Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information


You can suppress this exception and fall back to eager by setting:
    import torch._dynamo
    torch._dynamo.config.suppress_errors = True

2024-11-05 23:42:21,329 - root - ERROR - Traceback (most recent call last):
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1116, in visit_Call
    return fn(*args, **extra_kwargs, **kws)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\core.py", line 35, in wrapper
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\core.py", line 993, in to
    return semantic.cast(self, dtype, _builder, fp_downcast_rounding)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\semantic.py", line 759, in cast
    assert builder.options.allow_fp8e4nv, "fp8e4nv data type is not supported on CUDA arch < 89"
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
AssertionError: fp8e4nv data type is not supported on CUDA arch < 89

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1446, in _call_user_compiler
    compiled_fn = compiler_fn(gm, self.example_inputs())
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\repro\after_dynamo.py", line 129, in __call__
    compiled_gm = compiler_fn(gm, example_inputs)
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\__init__.py", line 2234, in __call__
    return compile_fx(model_, inputs_, config_patches=self.config)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 1521, in compile_fx
    return aot_autograd(
           ^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\backends\common.py", line 72, in __call__
    cg = aot_module_simplified(gm, example_inputs, **self.kwargs)
         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\aot_autograd.py", line 1071, in aot_module_simplified
    compiled_fn = dispatch_and_compile()
                  ^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\aot_autograd.py", line 1056, in dispatch_and_compile
    compiled_fn, _ = create_aot_dispatcher_function(
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\aot_autograd.py", line 522, in create_aot_dispatcher_function
    return _create_aot_dispatcher_function(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\aot_autograd.py", line 759, in _create_aot_dispatcher_function
    compiled_fn, fw_metadata = compiler_fn(
                               ^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\_aot_autograd\jit_compile_runtime_wrappers.py", line 179, in aot_dispatch_base
    compiled_fw = compiler(fw_module, updated_flat_args)
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 1350, in fw_compiler_base
    return _fw_compiler_base(model, example_inputs, is_inference)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 1421, in _fw_compiler_base
    return inner_compile(
           ^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 475, in compile_fx_inner
    return wrap_compiler_debug(_compile_fx_inner, compiler_name="inductor")(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\repro\after_aot.py", line 85, in debug_wrapper
    inner_compiled_fn = compiler_fn(gm, example_inputs)
                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 661, in _compile_fx_inner
    compiled_graph = FxGraphCache.load(
                     ^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\codecache.py", line 1334, in load
    compiled_graph = compile_fx_fn(
                     ^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 570, in codegen_and_compile
    compiled_graph = fx_codegen_and_compile(gm, example_inputs, **fx_kwargs)
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 878, in fx_codegen_and_compile
    compiled_fn = graph.compile_to_fn()
                  ^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\graph.py", line 1913, in compile_to_fn
    return self.compile_to_module().call
           ^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\graph.py", line 1839, in compile_to_module
    return self._compile_to_module()
           ^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\graph.py", line 1867, in _compile_to_module
    mod = PyCodeCache.load_by_key_path(
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\codecache.py", line 2876, in load_by_key_path
    mod = _reload_python_module(key, path)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\runtime\compile_tasks.py", line 45, in _reload_python_module
    exec(code, mod.__dict__, mod.__dict__)
  File "C:\Users\dante\AppData\Local\Temp\torchinductor_dante\23\c23gd6xwjyvqtoam26qkcrz4f7rkgmvxku35tah46cl4arkk5tbn.py", line 104, in <module>
    triton_per_fused_1 = async_compile.triton('triton_per_fused_1', '''
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\async_compile.py", line 203, in triton
    kernel.precompile()
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\runtime\triton_heuristics.py", line 244, in precompile
    compiled_binary, launcher = self._precompile_config(
                                ^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\runtime\triton_heuristics.py", line 443, in _precompile_config
    binary = triton.compile(*compile_args, **compile_kwargs)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\compiler.py", line 280, in compile
    module = src.make_ir(options, codegen_fns, context)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\compiler.py", line 113, in make_ir
    return ast_to_ttir(self.fn, self, context=context, options=options, codegen_fns=codegen_fns)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1297, in ast_to_ttir
    generator.visit(fn.parse())
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 359, in visit_Module
    ast.NodeVisitor.generic_visit(self, node)
  File "ast.py", line 426, in generic_visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 443, in visit_FunctionDef
    self.visit_compound_statement(node.body)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 351, in visit_compound_statement
    self.visit(stmt)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 496, in visit_Assign
    values = self.visit(node.value)
             ^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1124, in visit_Call
    raise CompilationError(self.jit_fn.src, node, None) from e
triton.compiler.errors.CompilationError: at 18:11:
    xindex = tl.full([1], xoffset, tl.int32)
    xmask = tl.full([RBLOCK], True, tl.int1)
    rindex = tl.arange(0, RBLOCK)[:]
    roffset = 0
    rmask = tl.full([RBLOCK], True, tl.int1)
    r1 = rindex
    x0 = xindex
    tmp0 = tl.load(in_ptr0 + (r1), None, eviction_policy='evict_last')
    tmp3 = tl.load(in_ptr1 + (r1 + (256*x0)), None)
    tmp1 = tmp0.to(tl.float32)
    tmp2 = tmp1.to(tl.float32)
    tmp4 = tmp3.to(tl.float32)
           ^

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
    process_inputs(input_dict, i)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1442, in sample
    return common_ksampler(model, seed, steps, cfg, sampler_name, scheduler, positive, negative, latent_image, denoise=denoise)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1409, in common_ksampler
    samples = comfy.sample.sample(model, noise, steps, cfg, sampler_name, scheduler, positive, negative, latent_image,
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\sample_error_enhancer.py", line 22, in informative_sample
    raise e
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\sample_error_enhancer.py", line 9, in informative_sample
    return original_sample(*args, **kwargs)  # This code helps interpret error messages that occur within exceptions but does not have any impact on other operations.
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\sample.py", line 43, in sample
    samples = sampler.sample(noise, positive, negative, cfg=cfg, latent_image=latent_image, start_step=start_step, last_step=last_step, force_full_denoise=force_full_denoise, denoise_mask=noise_mask, sigmas=sigmas, callback=callback, disable_pbar=disable_pbar, seed=seed)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 855, in sample
    return sample(self.model, noise, positive, negative, cfg, self.device, sampler, sigmas, self.model_options, latent_image=latent_image, denoise_mask=denoise_mask, callback=callback, disable_pbar=disable_pbar, seed=seed)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 753, in sample
    return cfg_guider.sample(noise, latent_image, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 740, in sample
    output = self.inner_sample(noise, latent_image, device, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 719, in inner_sample
    samples = sampler.sample(self, sigmas, extra_args, callback, noise, latent_image, denoise_mask, disable_pbar)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 624, in sample
    samples = self.sampler_function(model_k, noise, sigmas, extra_args=extra_args, callback=k_callback, disable=disable_pbar, **self.extra_options)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\k_diffusion\sampling.py", line 155, in sample_euler
    denoised = model(x, sigma_hat * s_in, **extra_args)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 299, in __call__
    out = self.inner_model(x, sigma, model_options=model_options, seed=seed)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 706, in __call__
    return self.predict_noise(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 709, in predict_noise
    return sampling_function(self.inner_model, x, timestep, self.conds.get("negative", None), self.conds.get("positive", None), self.cfg, model_options=model_options, seed=seed)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 279, in sampling_function
    out = calc_cond_batch(model, conds, x, timestep, model_options)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 228, in calc_cond_batch
    output = model.apply_model(input_x, timestep_, **c).chunk(batch_chunks)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\model_base.py", line 144, in apply_model
    model_output = self.diffusion_model(xc, t, context=context, control=control, transformer_options=transformer_options, **extra_conds).float()
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\eval_frame.py", line 465, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 1269, in __call__
    return self._torchdynamo_orig_callable(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 1064, in __call__
    result = self._inner_convert(
             ^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 526, in __call__
    return _compile(
           ^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 924, in _compile
    guarded_code = compile_inner(code, one_graph, hooks, transform)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 666, in compile_inner
    return _compile_inner(code, one_graph, hooks, transform)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_utils_internal.py", line 87, in wrapper_function
    return function(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 699, in _compile_inner
    out_code = transform_code_object(code, transform)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\bytecode_transformation.py", line 1322, in transform_code_object
    transformations(instructions, code_options)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 219, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 634, in transform
    tracer.run()
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 2796, in run
    super().run()
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 983, in run
    while self.step():
          ^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 895, in step
    self.dispatch_table[inst.opcode](self, inst)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 2987, in RETURN_VALUE
    self._return(inst)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 2972, in _return
    self.output.compile_subgraph(
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1117, in compile_subgraph
    self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1369, in compile_and_call_fx_graph
    compiled_fn = self.call_user_compiler(gm)
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1416, in call_user_compiler
    return self._call_user_compiler(gm)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1465, in _call_user_compiler
    raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: backend='inductor' raised:
CompilationError: at 18:11:
    xindex = tl.full([1], xoffset, tl.int32)
    xmask = tl.full([RBLOCK], True, tl.int1)
    rindex = tl.arange(0, RBLOCK)[:]
    roffset = 0
    rmask = tl.full([RBLOCK], True, tl.int1)
    r1 = rindex
    x0 = xindex
    tmp0 = tl.load(in_ptr0 + (r1), None, eviction_policy='evict_last')
    tmp3 = tl.load(in_ptr1 + (r1 + (256*x0)), None)
    tmp1 = tmp0.to(tl.float32)
    tmp2 = tmp1.to(tl.float32)
    tmp4 = tmp3.to(tl.float32)
           ^

Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information


You can suppress this exception and fall back to eager by setting:
    import torch._dynamo
    torch._dynamo.config.suppress_errors = True


2024-11-05 23:42:21,333 - root - INFO - Prompt executed in 54.52 seconds
2024-11-05 23:42:50,333 - root - INFO - got prompt
2024-11-05 23:42:51,106 - root - WARNING - Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
2024-11-05 23:42:51,122 - root - INFO - Requested to load Flux
2024-11-05 23:42:51,123 - root - INFO - Loading 1 new model
2024-11-05 23:42:55,882 - root - INFO - loaded completely 0.0 11350.048889160156 True
2024-11-05 23:42:57,135 - root - ERROR - !!! Exception during processing !!! backend='inductor' raised:
CompilationError: at 16:15:
    xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
    xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1)
    rbase = tl.arange(0, RBLOCK)[None, :]
    x0 = xindex
    _tmp11 = tl.full([XBLOCK, RBLOCK], 0, tl.float32)
    _tmp18 = tl.full([XBLOCK, RBLOCK], 0, tl.float32)
    for roffset in range(0, rnumel, RBLOCK):
        rindex = roffset + rbase
        rmask = rindex < rnumel
        r1 = rindex
        tmp0 = tl.load(in_ptr0 + (r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
        tmp6 = tl.load(in_ptr1 + (r1 + (3072*x0)), rmask, eviction_policy='evict_first', other=0.0)
               ^

Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information


You can suppress this exception and fall back to eager by setting:
    import torch._dynamo
    torch._dynamo.config.suppress_errors = True

2024-11-05 23:42:57,140 - root - ERROR - Traceback (most recent call last):
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1116, in visit_Call
    return fn(*args, **extra_kwargs, **kws)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\core.py", line 35, in wrapper
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\core.py", line 1597, in load
    return semantic.load(pointer, mask, other, boundary_check, padding_option, cache_modifier, eviction_policy,
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\semantic.py", line 1037, in load
    return _load_legacy(ptr, mask, other, boundary_check, padding, cache, eviction, is_volatile, builder)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\semantic.py", line 1005, in _load_legacy
    other = cast(other, elt_ty, builder)
            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\semantic.py", line 759, in cast
    assert builder.options.allow_fp8e4nv, "fp8e4nv data type is not supported on CUDA arch < 89"
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
AssertionError: fp8e4nv data type is not supported on CUDA arch < 89

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1446, in _call_user_compiler
    compiled_fn = compiler_fn(gm, self.example_inputs())
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\repro\after_dynamo.py", line 129, in __call__
    compiled_gm = compiler_fn(gm, example_inputs)
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\__init__.py", line 2234, in __call__
    return compile_fx(model_, inputs_, config_patches=self.config)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 1521, in compile_fx
    return aot_autograd(
           ^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\backends\common.py", line 72, in __call__
    cg = aot_module_simplified(gm, example_inputs, **self.kwargs)
         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\aot_autograd.py", line 1071, in aot_module_simplified
    compiled_fn = dispatch_and_compile()
                  ^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\aot_autograd.py", line 1056, in dispatch_and_compile
    compiled_fn, _ = create_aot_dispatcher_function(
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\aot_autograd.py", line 522, in create_aot_dispatcher_function
    return _create_aot_dispatcher_function(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\aot_autograd.py", line 759, in _create_aot_dispatcher_function
    compiled_fn, fw_metadata = compiler_fn(
                               ^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\_aot_autograd\jit_compile_runtime_wrappers.py", line 179, in aot_dispatch_base
    compiled_fw = compiler(fw_module, updated_flat_args)
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 1350, in fw_compiler_base
    return _fw_compiler_base(model, example_inputs, is_inference)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 1421, in _fw_compiler_base
    return inner_compile(
           ^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 475, in compile_fx_inner
    return wrap_compiler_debug(_compile_fx_inner, compiler_name="inductor")(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\repro\after_aot.py", line 85, in debug_wrapper
    inner_compiled_fn = compiler_fn(gm, example_inputs)
                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 661, in _compile_fx_inner
    compiled_graph = FxGraphCache.load(
                     ^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\codecache.py", line 1334, in load
    compiled_graph = compile_fx_fn(
                     ^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 570, in codegen_and_compile
    compiled_graph = fx_codegen_and_compile(gm, example_inputs, **fx_kwargs)
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 878, in fx_codegen_and_compile
    compiled_fn = graph.compile_to_fn()
                  ^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\graph.py", line 1913, in compile_to_fn
    return self.compile_to_module().call
           ^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\graph.py", line 1839, in compile_to_module
    return self._compile_to_module()
           ^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\graph.py", line 1867, in _compile_to_module
    mod = PyCodeCache.load_by_key_path(
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\codecache.py", line 2876, in load_by_key_path
    mod = _reload_python_module(key, path)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\runtime\compile_tasks.py", line 45, in _reload_python_module
    exec(code, mod.__dict__, mod.__dict__)
  File "C:\Users\dante\AppData\Local\Temp\torchinductor_dante\rv\crvvrv27opmilrlz3rk2536v747ahknp74h6ysrtxhjd4arakhb5.py", line 41, in <module>
    triton_red_fused_0 = async_compile.triton('triton_red_fused_0', '''
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\async_compile.py", line 203, in triton
    kernel.precompile()
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\runtime\triton_heuristics.py", line 244, in precompile
    compiled_binary, launcher = self._precompile_config(
                                ^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\runtime\triton_heuristics.py", line 443, in _precompile_config
    binary = triton.compile(*compile_args, **compile_kwargs)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\compiler.py", line 280, in compile
    module = src.make_ir(options, codegen_fns, context)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\compiler.py", line 113, in make_ir
    return ast_to_ttir(self.fn, self, context=context, options=options, codegen_fns=codegen_fns)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1297, in ast_to_ttir
    generator.visit(fn.parse())
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 359, in visit_Module
    ast.NodeVisitor.generic_visit(self, node)
  File "ast.py", line 426, in generic_visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 443, in visit_FunctionDef
    self.visit_compound_statement(node.body)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 351, in visit_compound_statement
    self.visit(stmt)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 964, in visit_For
    self.visit_compound_statement(node.body)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 351, in visit_compound_statement
    self.visit(stmt)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 496, in visit_Assign
    values = self.visit(node.value)
             ^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1124, in visit_Call
    raise CompilationError(self.jit_fn.src, node, None) from e
triton.compiler.errors.CompilationError: at 16:15:
    xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
    xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1)
    rbase = tl.arange(0, RBLOCK)[None, :]
    x0 = xindex
    _tmp11 = tl.full([XBLOCK, RBLOCK], 0, tl.float32)
    _tmp18 = tl.full([XBLOCK, RBLOCK], 0, tl.float32)
    for roffset in range(0, rnumel, RBLOCK):
        rindex = roffset + rbase
        rmask = rindex < rnumel
        r1 = rindex
        tmp0 = tl.load(in_ptr0 + (r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
        tmp6 = tl.load(in_ptr1 + (r1 + (3072*x0)), rmask, eviction_policy='evict_first', other=0.0)
               ^

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
    process_inputs(input_dict, i)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1442, in sample
    return common_ksampler(model, seed, steps, cfg, sampler_name, scheduler, positive, negative, latent_image, denoise=denoise)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1409, in common_ksampler
    samples = comfy.sample.sample(model, noise, steps, cfg, sampler_name, scheduler, positive, negative, latent_image,
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\sample_error_enhancer.py", line 22, in informative_sample
    raise e
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\sample_error_enhancer.py", line 9, in informative_sample
    return original_sample(*args, **kwargs)  # This code helps interpret error messages that occur within exceptions but does not have any impact on other operations.
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\sample.py", line 43, in sample
    samples = sampler.sample(noise, positive, negative, cfg=cfg, latent_image=latent_image, start_step=start_step, last_step=last_step, force_full_denoise=force_full_denoise, denoise_mask=noise_mask, sigmas=sigmas, callback=callback, disable_pbar=disable_pbar, seed=seed)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 855, in sample
    return sample(self.model, noise, positive, negative, cfg, self.device, sampler, sigmas, self.model_options, latent_image=latent_image, denoise_mask=denoise_mask, callback=callback, disable_pbar=disable_pbar, seed=seed)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 753, in sample
    return cfg_guider.sample(noise, latent_image, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 740, in sample
    output = self.inner_sample(noise, latent_image, device, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 719, in inner_sample
    samples = sampler.sample(self, sigmas, extra_args, callback, noise, latent_image, denoise_mask, disable_pbar)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 624, in sample
    samples = self.sampler_function(model_k, noise, sigmas, extra_args=extra_args, callback=k_callback, disable=disable_pbar, **self.extra_options)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\k_diffusion\sampling.py", line 155, in sample_euler
    denoised = model(x, sigma_hat * s_in, **extra_args)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 299, in __call__
    out = self.inner_model(x, sigma, model_options=model_options, seed=seed)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 706, in __call__
    return self.predict_noise(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 709, in predict_noise
    return sampling_function(self.inner_model, x, timestep, self.conds.get("negative", None), self.conds.get("positive", None), self.cfg, model_options=model_options, seed=seed)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 279, in sampling_function
    out = calc_cond_batch(model, conds, x, timestep, model_options)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 228, in calc_cond_batch
    output = model.apply_model(input_x, timestep_, **c).chunk(batch_chunks)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\model_base.py", line 144, in apply_model
    model_output = self.diffusion_model(xc, t, context=context, control=control, transformer_options=transformer_options, **extra_conds).float()
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\ldm\flux\model.py", line 159, in forward
    out = self.forward_orig(img, img_ids, context, txt_ids, timestep, y, guidance, control)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\ldm\flux\model.py", line 118, in forward_orig
    img, txt = block(img=img, txt=txt, vec=vec, pe=pe)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\eval_frame.py", line 465, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 1269, in __call__
    return self._torchdynamo_orig_callable(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 1064, in __call__
    result = self._inner_convert(
             ^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 526, in __call__
    return _compile(
           ^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 924, in _compile
    guarded_code = compile_inner(code, one_graph, hooks, transform)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 666, in compile_inner
    return _compile_inner(code, one_graph, hooks, transform)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_utils_internal.py", line 87, in wrapper_function
    return function(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 699, in _compile_inner
    out_code = transform_code_object(code, transform)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\bytecode_transformation.py", line 1322, in transform_code_object
    transformations(instructions, code_options)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 219, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 634, in transform
    tracer.run()
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 2796, in run
    super().run()
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 983, in run
    while self.step():
          ^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 895, in step
    self.dispatch_table[inst.opcode](self, inst)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 2987, in RETURN_VALUE
    self._return(inst)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 2972, in _return
    self.output.compile_subgraph(
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1142, in compile_subgraph
    self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1369, in compile_and_call_fx_graph
    compiled_fn = self.call_user_compiler(gm)
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1416, in call_user_compiler
    return self._call_user_compiler(gm)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1465, in _call_user_compiler
    raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: backend='inductor' raised:
CompilationError: at 16:15:
    xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
    xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1)
    rbase = tl.arange(0, RBLOCK)[None, :]
    x0 = xindex
    _tmp11 = tl.full([XBLOCK, RBLOCK], 0, tl.float32)
    _tmp18 = tl.full([XBLOCK, RBLOCK], 0, tl.float32)
    for roffset in range(0, rnumel, RBLOCK):
        rindex = roffset + rbase
        rmask = rindex < rnumel
        r1 = rindex
        tmp0 = tl.load(in_ptr0 + (r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
        tmp6 = tl.load(in_ptr1 + (r1 + (3072*x0)), rmask, eviction_policy='evict_first', other=0.0)
               ^

Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information


You can suppress this exception and fall back to eager by setting:
    import torch._dynamo
    torch._dynamo.config.suppress_errors = True


2024-11-05 23:42:57,144 - root - INFO - Prompt executed in 6.80 seconds
2024-11-05 23:43:30,905 - root - INFO - got prompt
2024-11-05 23:43:30,927 - root - WARNING - Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
2024-11-05 23:43:32,091 - root - ERROR - !!! Exception during processing !!! backend='inductor' raised:
CompilationError: at 16:15:
    xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
    xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1)
    rbase = tl.arange(0, RBLOCK)[None, :]
    x0 = xindex
    _tmp11 = tl.full([XBLOCK, RBLOCK], 0, tl.float32)
    _tmp18 = tl.full([XBLOCK, RBLOCK], 0, tl.float32)
    for roffset in range(0, rnumel, RBLOCK):
        rindex = roffset + rbase
        rmask = rindex < rnumel
        r1 = rindex
        tmp0 = tl.load(in_ptr0 + (r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
        tmp6 = tl.load(in_ptr1 + (r1 + (3072*x0)), rmask, eviction_policy='evict_first', other=0.0)
               ^

Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information


You can suppress this exception and fall back to eager by setting:
    import torch._dynamo
    torch._dynamo.config.suppress_errors = True

2024-11-05 23:43:32,096 - root - ERROR - Traceback (most recent call last):
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1116, in visit_Call
    return fn(*args, **extra_kwargs, **kws)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\core.py", line 35, in wrapper
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\core.py", line 1597, in load
    return semantic.load(pointer, mask, other, boundary_check, padding_option, cache_modifier, eviction_policy,
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\semantic.py", line 1037, in load
    return _load_legacy(ptr, mask, other, boundary_check, padding, cache, eviction, is_volatile, builder)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\semantic.py", line 1005, in _load_legacy
    other = cast(other, elt_ty, builder)
            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\language\semantic.py", line 759, in cast
    assert builder.options.allow_fp8e4nv, "fp8e4nv data type is not supported on CUDA arch < 89"
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
AssertionError: fp8e4nv data type is not supported on CUDA arch < 89

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1446, in _call_user_compiler
    compiled_fn = compiler_fn(gm, self.example_inputs())
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\repro\after_dynamo.py", line 129, in __call__
    compiled_gm = compiler_fn(gm, example_inputs)
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\__init__.py", line 2234, in __call__
    return compile_fx(model_, inputs_, config_patches=self.config)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 1521, in compile_fx
    return aot_autograd(
           ^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\backends\common.py", line 72, in __call__
    cg = aot_module_simplified(gm, example_inputs, **self.kwargs)
         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\aot_autograd.py", line 1071, in aot_module_simplified
    compiled_fn = dispatch_and_compile()
                  ^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\aot_autograd.py", line 1056, in dispatch_and_compile
    compiled_fn, _ = create_aot_dispatcher_function(
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\aot_autograd.py", line 522, in create_aot_dispatcher_function
    return _create_aot_dispatcher_function(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\aot_autograd.py", line 759, in _create_aot_dispatcher_function
    compiled_fn, fw_metadata = compiler_fn(
                               ^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_functorch\_aot_autograd\jit_compile_runtime_wrappers.py", line 179, in aot_dispatch_base
    compiled_fw = compiler(fw_module, updated_flat_args)
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 1350, in fw_compiler_base
    return _fw_compiler_base(model, example_inputs, is_inference)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 1421, in _fw_compiler_base
    return inner_compile(
           ^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 475, in compile_fx_inner
    return wrap_compiler_debug(_compile_fx_inner, compiler_name="inductor")(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\repro\after_aot.py", line 85, in debug_wrapper
    inner_compiled_fn = compiler_fn(gm, example_inputs)
                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 661, in _compile_fx_inner
    compiled_graph = FxGraphCache.load(
                     ^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\codecache.py", line 1334, in load
    compiled_graph = compile_fx_fn(
                     ^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 570, in codegen_and_compile
    compiled_graph = fx_codegen_and_compile(gm, example_inputs, **fx_kwargs)
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\compile_fx.py", line 878, in fx_codegen_and_compile
    compiled_fn = graph.compile_to_fn()
                  ^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\graph.py", line 1913, in compile_to_fn
    return self.compile_to_module().call
           ^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\graph.py", line 1839, in compile_to_module
    return self._compile_to_module()
           ^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\graph.py", line 1867, in _compile_to_module
    mod = PyCodeCache.load_by_key_path(
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\codecache.py", line 2876, in load_by_key_path
    mod = _reload_python_module(key, path)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\runtime\compile_tasks.py", line 45, in _reload_python_module
    exec(code, mod.__dict__, mod.__dict__)
  File "C:\Users\dante\AppData\Local\Temp\torchinductor_dante\63\c63xihbvnti4s2jyxtepls7wajkdjdhsqw47zmgin4vppxczfful.py", line 41, in <module>
    triton_red_fused_0 = async_compile.triton('triton_red_fused_0', '''
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\async_compile.py", line 203, in triton
    kernel.precompile()
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\runtime\triton_heuristics.py", line 244, in precompile
    compiled_binary, launcher = self._precompile_config(
                                ^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_inductor\runtime\triton_heuristics.py", line 443, in _precompile_config
    binary = triton.compile(*compile_args, **compile_kwargs)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\compiler.py", line 280, in compile
    module = src.make_ir(options, codegen_fns, context)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\compiler.py", line 113, in make_ir
    return ast_to_ttir(self.fn, self, context=context, options=options, codegen_fns=codegen_fns)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1297, in ast_to_ttir
    generator.visit(fn.parse())
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 359, in visit_Module
    ast.NodeVisitor.generic_visit(self, node)
  File "ast.py", line 426, in generic_visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 443, in visit_FunctionDef
    self.visit_compound_statement(node.body)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 351, in visit_compound_statement
    self.visit(stmt)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 964, in visit_For
    self.visit_compound_statement(node.body)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 351, in visit_compound_statement
    self.visit(stmt)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 496, in visit_Assign
    values = self.visit(node.value)
             ^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1204, in visit
    ret = super().visit(node)
          ^^^^^^^^^^^^^^^^^^^
  File "ast.py", line 418, in visit
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\compiler\code_generator.py", line 1124, in visit_Call
    raise CompilationError(self.jit_fn.src, node, None) from e
triton.compiler.errors.CompilationError: at 16:15:
    xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
    xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1)
    rbase = tl.arange(0, RBLOCK)[None, :]
    x0 = xindex
    _tmp11 = tl.full([XBLOCK, RBLOCK], 0, tl.float32)
    _tmp18 = tl.full([XBLOCK, RBLOCK], 0, tl.float32)
    for roffset in range(0, rnumel, RBLOCK):
        rindex = roffset + rbase
        rmask = rindex < rnumel
        r1 = rindex
        tmp0 = tl.load(in_ptr0 + (r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
        tmp6 = tl.load(in_ptr1 + (r1 + (3072*x0)), rmask, eviction_policy='evict_first', other=0.0)
               ^

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
    process_inputs(input_dict, i)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1442, in sample
    return common_ksampler(model, seed, steps, cfg, sampler_name, scheduler, positive, negative, latent_image, denoise=denoise)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1409, in common_ksampler
    samples = comfy.sample.sample(model, noise, steps, cfg, sampler_name, scheduler, positive, negative, latent_image,
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\sample_error_enhancer.py", line 22, in informative_sample
    raise e
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\sample_error_enhancer.py", line 9, in informative_sample
    return original_sample(*args, **kwargs)  # This code helps interpret error messages that occur within exceptions but does not have any impact on other operations.
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\sample.py", line 43, in sample
    samples = sampler.sample(noise, positive, negative, cfg=cfg, latent_image=latent_image, start_step=start_step, last_step=last_step, force_full_denoise=force_full_denoise, denoise_mask=noise_mask, sigmas=sigmas, callback=callback, disable_pbar=disable_pbar, seed=seed)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 855, in sample
    return sample(self.model, noise, positive, negative, cfg, self.device, sampler, sigmas, self.model_options, latent_image=latent_image, denoise_mask=denoise_mask, callback=callback, disable_pbar=disable_pbar, seed=seed)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 753, in sample
    return cfg_guider.sample(noise, latent_image, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 740, in sample
    output = self.inner_sample(noise, latent_image, device, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 719, in inner_sample
    samples = sampler.sample(self, sigmas, extra_args, callback, noise, latent_image, denoise_mask, disable_pbar)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 624, in sample
    samples = self.sampler_function(model_k, noise, sigmas, extra_args=extra_args, callback=k_callback, disable=disable_pbar, **self.extra_options)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\k_diffusion\sampling.py", line 155, in sample_euler
    denoised = model(x, sigma_hat * s_in, **extra_args)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 299, in __call__
    out = self.inner_model(x, sigma, model_options=model_options, seed=seed)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 706, in __call__
    return self.predict_noise(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 709, in predict_noise
    return sampling_function(self.inner_model, x, timestep, self.conds.get("negative", None), self.conds.get("positive", None), self.cfg, model_options=model_options, seed=seed)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 279, in sampling_function
    out = calc_cond_batch(model, conds, x, timestep, model_options)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 228, in calc_cond_batch
    output = model.apply_model(input_x, timestep_, **c).chunk(batch_chunks)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\model_base.py", line 144, in apply_model
    model_output = self.diffusion_model(xc, t, context=context, control=control, transformer_options=transformer_options, **extra_conds).float()
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\ldm\flux\model.py", line 159, in forward
    out = self.forward_orig(img, img_ids, context, txt_ids, timestep, y, guidance, control)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\ComfyUI\comfy\ldm\flux\model.py", line 118, in forward_orig
    img, txt = block(img=img, txt=txt, vec=vec, pe=pe)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\eval_frame.py", line 465, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 1269, in __call__
    return self._torchdynamo_orig_callable(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 1064, in __call__
    result = self._inner_convert(
             ^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 526, in __call__
    return _compile(
           ^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 924, in _compile
    guarded_code = compile_inner(code, one_graph, hooks, transform)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 666, in compile_inner
    return _compile_inner(code, one_graph, hooks, transform)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_utils_internal.py", line 87, in wrapper_function
    return function(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 699, in _compile_inner
    out_code = transform_code_object(code, transform)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\bytecode_transformation.py", line 1322, in transform_code_object
    transformations(instructions, code_options)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 219, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\convert_frame.py", line 634, in transform
    tracer.run()
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 2796, in run
    super().run()
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 983, in run
    while self.step():
          ^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 895, in step
    self.dispatch_table[inst.opcode](self, inst)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 2987, in RETURN_VALUE
    self._return(inst)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\symbolic_convert.py", line 2972, in _return
    self.output.compile_subgraph(
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1142, in compile_subgraph
    self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1369, in compile_and_call_fx_graph
    compiled_fn = self.call_user_compiler(gm)
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1416, in call_user_compiler
    return self._call_user_compiler(gm)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "C:\NeuroAI\Photo\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\output_graph.py", line 1465, in _call_user_compiler
    raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: backend='inductor' raised:
CompilationError: at 16:15:
    xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
    xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1)
    rbase = tl.arange(0, RBLOCK)[None, :]
    x0 = xindex
    _tmp11 = tl.full([XBLOCK, RBLOCK], 0, tl.float32)
    _tmp18 = tl.full([XBLOCK, RBLOCK], 0, tl.float32)
    for roffset in range(0, rnumel, RBLOCK):
        rindex = roffset + rbase
        rmask = rindex < rnumel
        r1 = rindex
        tmp0 = tl.load(in_ptr0 + (r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
        tmp6 = tl.load(in_ptr1 + (r1 + (3072*x0)), rmask, eviction_policy='evict_first', other=0.0)
               ^

Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information


You can suppress this exception and fall back to eager by setting:
    import torch._dynamo
    torch._dynamo.config.suppress_errors = True


2024-11-05 23:43:32,100 - root - INFO - Prompt executed in 1.19 seconds

Attached Workflow

Please make sure that workflow does not contain any sensitive information such as API keys or passwords.

Workflow too large. Please manually upload the workflow from local file system.

Additional Context

(Please add any additional context or steps to reproduce the error here)

@woct0rdho
Copy link
Owner Author

@Danteday It says fp8e4nv data type is not supported on CUDA arch < 89. Sadly, this is only supported on newer GPUs like 40xx

@Danteday
Copy link

Danteday commented Nov 6, 2024

@DantedayТам написано fp8e4nv data type is not supported on CUDA arch < 89. К сожалению, это поддерживается только на новых графических процессорах, таких как 40xx

It's a pity, I thought everything would work on my 3090 :C

@Dahvikiin
Copy link

@DantedayТам написано fp8e4nv data type is not supported on CUDA arch < 89. К сожалению, это поддерживается только на новых графических процессорах, таких как 40xx

It's a pity, I thought everything would work on my 3090 :C

I think that --fast is only for SM89 (RTX4000+)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests