Skip to content

Commit 2aead0e

Browse files
committed
fix kv cache apply
Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>
1 parent fafccf3 commit 2aead0e

File tree

1 file changed

+1
-1
lines changed
  • src/compressed_tensors/quantization/lifecycle

1 file changed

+1
-1
lines changed

src/compressed_tensors/quantization/lifecycle/apply.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -137,7 +137,7 @@ def apply_quantization_config(
137137
# because attention quantization is a superset of kv cache quantization
138138
if config.kv_cache_scheme is not None:
139139
scheme = QuantizationScheme(
140-
targets=".*self_attn$", input_activations=config.kv_cache_scheme
140+
targets=[".*self_attn$"], input_activations=config.kv_cache_scheme
141141
)
142142
for submodule in model.modules():
143143
if is_attention_module(submodule):

0 commit comments

Comments
 (0)