You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
f"Asked to export a {model_type} model for the task {task}{autodetected_message}, but the Optimum ONNX exporter only supports the tasks {', '.join(model_tasks.keys())} for {model_type}. Please use a supported task. Please open an issue at https://github.com/huggingface/optimum/issues if you would like the task {task} to be supported in the ONNX export for {model_type}."
384
387
)
@@ -422,7 +425,13 @@ def main_export(
422
425
"Could not infer the pad token id, which is needed in this case, please provide it with the --pad_token_id argument"
f"Trying to export a {model.config.model_type} model, that is a custom or unsupported architecture, but no custom onnx configuration was passed as `custom_onnx_configs`. Please refer to https://huggingface.co/docs/optimum/main/en/exporters/onnx/usage_guides/export_a_model#custom-export-of-transformers-models for an example on how to export custom models. Please open an issue at https://github.com/huggingface/optimum/issues if you would like the model type {model.config.model_type} to be supported natively in the ONNX export."
538
+
f"Trying to export a {model_type} model, that is a custom or unsupported architecture, but no custom onnx configuration was passed as `custom_onnx_configs`. Please refer to https://huggingface.co/docs/optimum/main/en/exporters/onnx/usage_guides/export_a_model#custom-export-of-transformers-models for an example on how to export custom models. Please open an issue at https://github.com/huggingface/optimum/issues if you would like the model type {model_type} to be supported natively in the ONNX export."
524
539
)
525
540
526
541
iftaskisNone:
@@ -690,7 +705,7 @@ def onnx_export(
690
705
iflibrary_name=="diffusers":
691
706
# TODO: fix Can't pickle local object 'get_stable_diffusion_models_for_export.<locals>.<lambda>'
692
707
use_subprocess=False
693
-
elifmodel.config.model_typeinUNPICKABLE_ARCHS:
708
+
elifmodel_typeinUNPICKABLE_ARCHS:
694
709
# Pickling is bugged for nn.utils.weight_norm: https://github.com/pytorch/pytorch/issues/102983
695
710
# TODO: fix "Cowardly refusing to serialize non-leaf tensor" error for wav2vec2-conformer
0 commit comments