Skip to content

Commit c2bce52

Browse files
committed
fix typos; remove comment
1 parent e5864f8 commit c2bce52

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

vllm/model_executor/layers/quantization/compressed_tensors/compressed_tensors_moe.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -272,7 +272,7 @@ def create_weights(self, layer: torch.nn.Module, num_experts: int,
272272
params_dtype: torch.dtype, **extra_weight_attrs):
273273

274274
assert params_dtype == torch.float16, (
275-
"float16 is required for MoE compressd models. Set dtype=torch.float16" # noqa: E501
275+
"float16 is required for MoE compressed models. Set dtype=torch.float16" # noqa: E501
276276
)
277277

278278
# Will transpose the loaded weight along the
@@ -306,7 +306,7 @@ def create_weights(self, layer: torch.nn.Module, num_experts: int,
306306
load_full_w2 = self.actorder and self.group_size != -1
307307
w2_scales_size = (intermediate_full
308308
if load_full_w2 else intermediate_size)
309-
# @eliza TODO: is this condition actually needed/is it doing anything?
309+
310310
self.is_k_full = (not self.actorder) or (intermediate_size
311311
== intermediate_full)
312312

vllm/model_executor/layers/quantization/compressed_tensors/schemes/compressed_tensors_w4a16_24.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -62,7 +62,7 @@ def create_weights(self, layer: torch.nn.Module, input_size: int,
6262
**kwargs):
6363

6464
assert params_dtype == torch.float16, (
65-
"float16 is required for marlin24 compressd models. Set dtype=torch.float16" # noqa: E501
65+
"float16 is required for marlin24 compressed models. Set dtype=torch.float16" # noqa: E501
6666
)
6767

6868
pack_factor = 32 // self.quant_type.size_bits

0 commit comments

Comments
 (0)