3333 float8_dynamic_activation_float8_weight ,
3434 float8_static_activation_float8_weight ,
3535 float8_weight_only ,
36+ int4_dynamic_activation_int4_weight ,
3637 int4_weight_only ,
3738 int8_dynamic_activation_int4_weight ,
3839 int8_dynamic_activation_int8_weight ,
5051 TORCH_VERSION_AT_LEAST_2_5 ,
5152 TORCH_VERSION_AT_LEAST_2_6 ,
5253 is_sm_at_least_89 ,
54+ is_sm_at_least_90 ,
5355 unwrap_tensor_subclass ,
5456)
5557
@@ -798,6 +800,10 @@ def test_int4wo_cpu(self, dtype, x_dim):
798800 float8_weight_only (),
799801 float8_dynamic_activation_float8_weight (),
800802 float8_static_activation_float8_weight (scale = torch .tensor ([1.0 ])),
803+ int4_dynamic_activation_int4_weight (),
804+ int8_dynamic_activation_int8_weight (),
805+ int8_dynamic_activation_int4_weight (),
806+ int8_weight_only (),
801807 ],
802808 )
803809 def test_workflow_e2e_numerics (self , config ):
@@ -816,6 +822,11 @@ def test_workflow_e2e_numerics(self, config):
816822 and not is_sm_at_least_89 ()
817823 ):
818824 return unittest .skip ("requires CUDA capability 8.9 or greater" )
825+ elif (
826+ isinstance (config , int4_dynamic_activation_int4_weight )
827+ and is_sm_at_least_90 ()
828+ ):
829+ return unittest .skip ("only supported on CUDA capability 8.9, not greater" )
819830
820831 # scale has to be moved to cuda here because the parametrization init
821832 # code happens before gating for cuda availability
0 commit comments