You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
However, when I try to reload the model, I get the following:
>>> AutoModelForCausalLM.from_pretrained("test_save")
File "/home/matt/PycharmProjects/transformers/src/transformers/modeling_utils.py", line 806, in _load_state_dict_into_meta_model
not hf_quantizer.check_quantized_param(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/matt/PycharmProjects/transformers/src/transformers/quantizers/quantizer_finegrained_fp8.py", line 155, in check_quantized_param
raise ValueError("Expect quantized weights but got an unquantized weight")
ValueError: Expect quantized weights but got an unquantized weight
It seems like even though we support FP8 loading after #36828, we may not be saving it correctly? cc @kylesayrs
The text was updated successfully, but these errors were encountered:
I tried making a "mini-Deepseek" for testing but encountered some issues. This works fine:
However, when I try to reload the model, I get the following:
It seems like even though we support FP8 loading after #36828, we may not be saving it correctly? cc @kylesayrs
The text was updated successfully, but these errors were encountered: