-
Notifications
You must be signed in to change notification settings - Fork 36
Pull requests: HabanaAI/vllm-hpu-extension
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Fix: Pass correct long_context flag to warmup_range_with_limit
#297
opened Jul 17, 2025 by
yafshar
Loading…
[SW-228042] Add support for dynamic vLLM kv-cache quantization
#292
opened Jul 16, 2025 by
dudilester
Loading…
Enable calibration using pile-10k dataset for DeepSeek models
#279
opened Jul 14, 2025 by
yangulei
Loading…
Fix for calibration error TypeError: generate_responses() missing 1 required positional argument: 'args'
#255
opened Jul 2, 2025 by
tthakkal
Loading…
Allow usage of fused_block_softmax_adjustment for Qwen with Lazy
#246
opened Jun 27, 2025 by
mswiniarsk
•
Draft
Use sets for faster filter checks. Better long context support
#203
opened May 28, 2025 by
pi314ever
Loading…
[SW-225565] Enable triangular softmax with merged prefill
#197
opened May 26, 2025 by
kamil-kaczor
•
Draft
Previous Next
ProTip!
Follow long discussions with comments:>50.