Skip to content
Merged
Changes from 6 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion src/transformers/integrations/finegrained_fp8.py
Original file line number Diff line number Diff line change
Expand Up @@ -606,7 +606,7 @@ def replace_with_fp8_linear(
module_kwargs = {} if pre_quantized else {"dtype": None}
new_module = None
with init_empty_weights():
if "gate_up_proj" in module_name or "down_proj" in module_name and "experts" in module_name:
if module_name.endswith(".experts"):
new_module = FP8Expert(
config=model.config, block_size=quantization_config.weight_block_size, **module_kwargs
)
Expand Down