We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 3edd804 commit 2a596f5Copy full SHA for 2a596f5
src/transformers/models/esm/modeling_esm.py
@@ -589,6 +589,7 @@ class EsmPreTrainedModel(PreTrainedModel):
589
config: EsmConfig
590
base_model_prefix = "esm"
591
supports_gradient_checkpointing = True
592
+ accepts_loss_kwargs = False
593
_no_split_modules = ["EsmLayer", "EsmFoldTriangularSelfAttentionBlock", "EsmEmbeddings"]
594
_keys_to_ignore_on_load_unexpected = ["position_embeddings.weight"]
595
_supports_flash_attn = True
0 commit comments