Skip to content

Commit 46a7b0c

Browse files
committed
Merge remote-tracking branch 'origin/fix_ci' into add_h100_ci
2 parents 6dd36b3 + 08bccdc commit 46a7b0c

File tree

2 files changed

+1
-9
lines changed

2 files changed

+1
-9
lines changed

test/integration/test_integration.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -1821,7 +1821,7 @@ def test_autoquant_int4wo(self, device, dtype):
18211821
self.assertGreater(compute_error(ref, out), 20)
18221822

18231823
@parameterized.expand(COMMON_DEVICE_DTYPE)
1824-
@unittest.skipIf(not torch.cuda.is_available(), "Need CUDA available")
1824+
@unittest.skipIf(not is_sm_at_least_90(), "Need cuda arch greater than SM90")
18251825
@unittest.skipIf(
18261826
not TORCH_VERSION_AT_LEAST_2_5, "autoquant int4 option requires 2.5+."
18271827
)

torchao/quantization/qat/utils.py

-8
Original file line numberDiff line numberDiff line change
@@ -16,14 +16,6 @@
1616
_get_per_token_block_size,
1717
)
1818

19-
# Attribute name representing the forward prehook wrapping the
20-
# linear input in an `AffineFakeQuantizedTensor` on a linear module.
21-
#
22-
# The value of this attribute is a 2-tuple of (prehook, handle).
23-
# The prehook can be disabled by calling `handle.remove()`, and
24-
# re-enabled by calling `module.register_forward_pre_hook(prehook)`.
25-
_QAT_LINEAR_SUBCLASS_INPUT_PREHOOK = "_qat_linear_subclass_input_prehook"
26-
2719

2820
class _GenericFakeQuantize(torch.autograd.Function):
2921
"""

0 commit comments

Comments
 (0)