Skip to content

Commit 51ff216

Browse files
author
George
authored
[Bugfix] update should_ignore_layer (vllm-project#11354)
Signed-off-by: George Ohashi <[email protected]>
1 parent dd2b563 commit 51ff216

File tree

1 file changed

+1
-1
lines changed
  • vllm/model_executor/layers/quantization/compressed_tensors

1 file changed

+1
-1
lines changed

vllm/model_executor/layers/quantization/compressed_tensors/utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ def should_ignore_layer(layer_name: Optional[str],
3030
# in the safetensors checkpoint. So, we convert the name
3131
# from the fused version to unfused + check to make sure that
3232
# each shard of the fused layer has the same scheme.
33-
if proj_name in FUSED_LAYER_NAME_MAPPING:
33+
if proj_name in FUSED_LAYER_NAME_MAPPING and layer_name not in ignore:
3434
shard_proj_names = FUSED_LAYER_NAME_MAPPING[proj_name]
3535

3636
# Convert fused_name --> [shard_names]

0 commit comments

Comments
 (0)