Skip to content

Commit fafccf3

Browse files
committed
add tests
Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>
1 parent 5229f86 commit fafccf3

File tree

2 files changed

+1
-4
lines changed

2 files changed

+1
-4
lines changed

src/compressed_tensors/modeling/attention.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -99,9 +99,7 @@ def _ct_hooked_attention(module: Module, *args, **kwargs):
9999
return ALL_ATTENTION_FUNCTIONS[_original_impl](module, *args, **kwargs)
100100

101101

102-
def initialize_hooked_attention(
103-
model: PreTrainedModel, module: Module, quantize: bool = True
104-
):
102+
def initialize_hooked_attention(model: PreTrainedModel, module: Module):
105103
"""
106104
Initialize `QuantizedAttentionImpl` and `QuantizedKVCache` instances
107105
attached to attention

src/compressed_tensors/quantization/lifecycle/apply.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,6 @@
2929
initialize_module_for_quantization,
3030
is_attention_module,
3131
)
32-
from compressed_tensors.quantization.quant_args import QuantizationArgs
3332
from compressed_tensors.quantization.quant_config import (
3433
QuantizationConfig,
3534
QuantizationStatus,

0 commit comments

Comments
 (0)