Skip to content
Discussion options

You must be logged in to vote

Okay, so I did a count on the number of parameters per layer type on this model:

  • torch.nn.modules.conv.Conv2d: 140772,
  • torch.nn.modules.linear.Linear: 461472,
  • torch.nn.modules.normalization.GroupNorm: 2848,
  • torch.nn.modules.normalization.LayerNorm: 1152,
  • torch.nn.modules.conv.Conv3d: 71488

So most of the parameters are on Conv2d and Linear, not on Conv3d (which is not supported), so in theory, using LoRA could be helpful.

This is also being trained from scratch with no pretraining

This is a big problem. PEFT is intended for fine-tuning, i.e. taking a pretrained model and adapting it to your specific problem. You will almost certainly not succeed when training from scratch. I imagine y…

Replies: 1 comment 6 replies

Comment options

You must be logged in to vote
6 replies
@RandomGamingDev
Comment options

@BenjaminBossan
Comment options

@RandomGamingDev
Comment options

@BenjaminBossan
Comment options

Answer selected by RandomGamingDev
@RandomGamingDev
Comment options

@BenjaminBossan
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants