You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
PS D:\ComfyUI_windows_portable_nvidia\ComfyUI_windows_portable> .\run_nvidia_gpu.bat
D:\ComfyUI_windows_portable_nvidia\ComfyUI_windows_portable>.\python_embeded\python.exe-s ComfyUI\main.py --windows-standalone-build
Checkpoint files will always be loaded safely.
Total VRAM 8191 MB, total RAM 16317 MB
pytorch version: 2.6.0+cu126
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 3060 Ti : cudaMallocAsync
Using pytorch attention
ComfyUI version: 0.3.19
[PromptServer] web root: D:\ComfyUI_windows_portable_nvidia\ComfyUI_windows_portable\python_embeded\Lib\site-packages\comfyui_frontend_package\static
Import times for custom nodes:
0.0 seconds: D:\ComfyUI_windows_portable_nvidia\ComfyUI_windows_portable\ComfyUI\custom_nodes\websocket_image_save.py
Starting server
To see the GUI go to: http://127.0.0.1:8188
got prompt
Using pytorch attention in VAE
Using pytorch attention in VAE
VAE load device: cuda:0, offload device: cpu, dtype: torch.bfloat16
D:\ComfyUI_windows_portable_nvidia\ComfyUI_windows_portable>pause
请按任意键继续.. .
Other
No response
The text was updated successfully, but these errors were encountered:
Your question
When I try flux, it paused
Logs
Other
No response
The text was updated successfully, but these errors were encountered: