Skip to content

Commit

Permalink
fix typos; remove comment
Browse files Browse the repository at this point in the history
  • Loading branch information
dsikka committed Jan 17, 2025
1 parent c0c5d7c commit a68004b
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 3 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -272,7 +272,7 @@ def create_weights(self, layer: torch.nn.Module, num_experts: int,
params_dtype: torch.dtype, **extra_weight_attrs):

assert params_dtype == torch.float16, (
"float16 is required for MoE compressd models. Set dtype=torch.float16" # noqa: E501
"float16 is required for MoE compressed models. Set dtype=torch.float16" # noqa: E501
)

# Will transpose the loaded weight along the
Expand Down Expand Up @@ -306,7 +306,7 @@ def create_weights(self, layer: torch.nn.Module, num_experts: int,
load_full_w2 = self.actorder and self.group_size != -1
w2_scales_size = (intermediate_full
if load_full_w2 else intermediate_size)
# @eliza TODO: is this condition actually needed/is it doing anything?

self.is_k_full = (not self.actorder) or (intermediate_size
== intermediate_full)

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,7 @@ def create_weights(self, layer: torch.nn.Module, input_size: int,
**kwargs):

assert params_dtype == torch.float16, (
"float16 is required for marlin24 compressd models. Set dtype=torch.float16" # noqa: E501
"float16 is required for marlin24 compressed models. Set dtype=torch.float16" # noqa: E501
)

pack_factor = 32 // self.quant_type.size_bits
Expand Down

0 comments on commit a68004b

Please sign in to comment.