LayerInfo doesn't support fp8 and int4_awq dtype? #2547
Labels
Investigating
Low Precision
Issue about lower bit quantization, including int8, int4, fp8
triaged
Issue has been triaged by maintainers
I built LLaMA 3B w4a8_awq by blow command.
and then, export layer information.
However, there is no Int4 description for
WeightOnlyGroupwiseQuantMatmul
weights.layerinfo.json
The text was updated successfully, but these errors were encountered: