[Bugfix] Enable teacahce in QwenImageEditPlusPipeline#379
Conversation
Signed-off-by: yuanheng <jonathan.zhaoyh@gmail.com>
Signed-off-by: yuanheng <jonathan.zhaoyh@gmail.com>
Signed-off-by: yuanheng <jonathan.zhaoyh@gmail.com>
good catch. I'll fix it later |
hsliuustc0106
left a comment
There was a problem hiding this comment.
lgtm, one question: does it apply to other qwen-image models as well or we need another a PR?
Hey @hsliuustc0106 , TeaCache has been enabled for the following Qwen-Image family pipelines, with the same extractor
It's expected to work for other Qwen family models which applies the same transformer forward flow. However, there might be some arg diff to adapt and test, like extra args imported into model and transformer block forward methods by Another note: As there exist many duplicate / similar code patterns for Qwen-Image-family pipelines, we might want to re-use some code and reduce redundancy, by applying a CacheMixin or a QwenImageBasePipeline (or some cleaner pattern), in the future. |
PTAL #395 |
Feel free to open a pr |
Signed-off-by: yuanheng <jonathan.zhaoyh@gmail.com> Signed-off-by: Didan Deng <33117903+wtomin@users.noreply.github.com>
Did a bit cleanup in #416 |
Signed-off-by: yuanheng <jonathan.zhaoyh@gmail.com> Signed-off-by: wangyu31577 <wangyu31577@hundsun.com>
Signed-off-by: yuanheng <jonathan.zhaoyh@gmail.com>
PLEASE FILL IN THE PR DESCRIPTION HERE ENSURING ALL CHECKLIST ITEMS (AT THE BOTTOM) HAVE BEEN CONSIDERED.
Purpose
Enable TeaCache in
QwenImageEditPlusPipelineTest Plan
Offline inference with running the example script examples/offline_inference/image_to_image/image_edit.py
Test Result
Tested on H100.
Total Generation Time
Input Images
Edited Images
Note: During testing the following error happened because of the warmup dummy run does not include an image path into the request. This won't affect the execution of actual request though.
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model.BEFORE SUBMITTING, PLEASE READ https://github.com/vllm-project/vllm-omni/blob/main/CONTRIBUTING.md (anything written below this line will be removed by GitHub Actions)