forked from YanWenKun/ComfyUI-Docker
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathREADME.txt
More file actions
104 lines (77 loc) · 3.03 KB
/
README.txt
File metadata and controls
104 lines (77 loc) · 3.03 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
Usage
-----
1. Check the Linux Kernel Module (driver) for Intel GPU:
lsmod | grep -i xe
For desktop distros with rolling-ish kernel updates, xe is usually included already.
2. Search and install `intel-compute-runtime` (or similar) with your host OS' package manager.
If not found, just ignore it and run directly.
For Ubuntu, follow this instruction and install the compute-related packages:
https://dgpu-docs.intel.com/driver/client/overview.html
3. Run the container:
====
mkdir -p \
storage \
storage-models/models \
storage-models/hf-hub \
storage-models/torch-hub \
storage-user/input \
storage-user/output \
storage-user/workflows
podman run -it --rm \
--name comfyui-xpu \
--device=/dev/dri \
--ipc=host \
--security-opt label=disable \
-p 8188:8188 \
-v "$(pwd)"/storage:/root \
-v "$(pwd)"/storage-models/models:/root/ComfyUI/models \
-v "$(pwd)"/storage-models/hf-hub:/root/.cache/huggingface/hub \
-v "$(pwd)"/storage-models/torch-hub:/root/.cache/torch/hub \
-v "$(pwd)"/storage-user/input:/root/ComfyUI/input \
-v "$(pwd)"/storage-user/output:/root/ComfyUI/output \
-v "$(pwd)"/storage-user/workflows:/root/ComfyUI/user/default/workflows \
-e CLI_ARGS="--disable-smart-memory --async-offload" \
yanwk/comfyui-boot:xpu
====
Note:
1. "--disable-smart-memory" is useful for XPU, although it consumes more RAM.
2. "--async-offload" gives subtle improvement and no visible harm.
More CLI_ARGS:
--lowvram
--bf16-unet
--bf16-vae
--bf16-text-enc
--mmap-torch-files
--reserve-vram 1
Check the doc before use:
https://github.com/comfyanonymous/ComfyUI/blob/master/comfy/cli_args.py
Test result on Arc B580
-----------------------
Time on generating one 1024x1024px image (pre-warmed, only infer time)
* SD 1.5 (20-step): OK (5s) | 512x512: GOOD (1s)
* SDXL (28-step): OK (6s)
* SDXL Refiner (20+5-step): OK (4+1s)
* Stable Cascade (20+10-step): FAST (3+3s)
* SD 3.5 large fp8 (20-step): GOOD (26s)
* Flux1 schnell (4-step): GOOD (8s)
* Flux1 dev fp8 (20-step): GOOD (35s)
* Flux1 Krea dev (20-step): OK (46s)
* Hunyuan3D 2.0 (Comfy Repackaged): 'Non-uniform work-groups are not supported by the target device'
* StableZero123: 'Non-uniform work-groups are not supported by the target device'
* Wan 2.2 5B Text to Video (640x352, 121 frames): FAST (60s, very poor quality)
* Wan 2.2 5B Text to Video (960x544, 121 frames): OK (181s, low quality)
* Wan 2.2 5B Text to Video (1280x704, 121 frames): FAILED (OOM)
Notes
--------------
1. Once OOM, ComfyUI won't crash, but GPU will stop responding
for future prompt (UR_RESULT_ERROR_DEVICE_LOST).
In that case, you will need to restart ComfyUI.
2. IPEX (Intel Extension for PyTorch) is not included, as most of its features
have been merged upstream and it is being phased out:
https://github.com/intel/intel-extension-for-pytorch/releases
For Windows Users
-----------------
Just use Intel AI Playground:
https://game.intel.com/us/stories/introducing-ai-playground/
Or yet another ComfyUI portable:
https://github.com/YanWenKun/ComfyUI-WinPortable-XPU