File tree Expand file tree Collapse file tree 2 files changed +7
-0
lines changed
test/prototype/safetensors
torchao/prototype/safetensors Expand file tree Collapse file tree 2 files changed +7
-0
lines changed Original file line number Diff line number Diff line change 2020from torchao .quantization .quant_api import (
2121 Float8DynamicActivationFloat8WeightConfig ,
2222 Int4WeightOnlyConfig ,
23+ Int8DynamicActivationIntxWeightConfig ,
24+ IntxWeightOnlyConfig ,
2325)
2426from torchao .utils import is_sm_at_least_89
2527
@@ -46,6 +48,8 @@ class TestSafeTensors(TestCase):
4648 (Int4WeightOnlyConfig (), False ),
4749 (Int4WeightOnlyConfig (), True ),
4850 (Int4WeightOnlyConfig (int4_packing_format = "tile_packed_to_4d" ), False ),
51+ (IntxWeightOnlyConfig (), False ),
52+ (Int8DynamicActivationIntxWeightConfig (), False ),
4953 ],
5054 )
5155 def test_safetensors (self , config , act_pre_scale = False ):
Original file line number Diff line number Diff line change 1010 Float8Tensor ,
1111 Int4Tensor ,
1212 Int4TilePackedTo4dTensor ,
13+ IntxUnpackedToInt8Tensor ,
1314)
1415from torchao .quantization .quantize_ .common import KernelPreference
1516from torchao .quantization .quantize_ .workflows import QuantizeTensorToFloat8Kwargs
1819 "Float8Tensor" : Float8Tensor ,
1920 "Int4Tensor" : Int4Tensor ,
2021 "Int4TilePackedTo4dTensor" : Int4TilePackedTo4dTensor ,
22+ "IntxUnpackedToInt8Tensor" : IntxUnpackedToInt8Tensor ,
2123 "Float8MMConfig" : torchao .float8 .inference .Float8MMConfig ,
2224 "QuantizeTensorToFloat8Kwargs" : QuantizeTensorToFloat8Kwargs ,
2325 "PerRow" : torchao .quantization .PerRow ,
2931 "Float8Tensor" ,
3032 "Int4Tensor" ,
3133 "Int4TilePackedTo4dTensor" ,
34+ "IntxUnpackedToInt8Tensor" ,
3235]
3336
3437__all__ = [
You can’t perform that action at this time.
0 commit comments