We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent f60594f commit e7b6f14Copy full SHA for e7b6f14
torchchat/utils/gguf_loader.py
@@ -237,7 +237,7 @@ def _prepare_weight_and_scales_and_zeros(
237
weight_int32, scales_and_zeros = group_quantize_tensor(
238
weight_bf16, n_bit=4, groupsize=groupsize
239
)
240
- if is_device(weight_int32.device.type, "cpu") and TORCH_VERSION_AT_LEAST_2_6:
+ if is_device(weight_int32.device.type, "cpu"):
241
weight_int4pack = torch.ops.aten._convert_weight_to_int4pack_for_cpu(
242
weight_int32, inner_k_tiles
243
0 commit comments