Skip to content

INT8 inference issues w8a8 #4

@meven3000

Description

@meven3000

Hi am am trying to use the Zentorch modules via fp32 with in8 for w8a8, currently i've tried using Quark compile Mistral 7b 0.2 and now also PytorchAO compiled W8a8.

Pytorch W8a8 works as expected without ZenTorch.

Tested from PyTorch ans associated ZenTorch version 2.4 -> 2.8

PyTorch version 2 7 0 (w8a8).txt

basic in8 script.txt

See attached for event log and associated basic testing script.

Noting: assuming again this may be due to enforced AVX512 for int8 (which probably should not be in this case), but certainly no mention of it this time.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions