Skip to content

Commit 5ee47ce

Browse files
authored
DOC Update DeLoRA constraints (#2854)
1 parent 8a008d6 commit 5ee47ce

File tree

1 file changed

+8
-3
lines changed

1 file changed

+8
-3
lines changed

docs/source/package_reference/delora.md

Lines changed: 8 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -18,9 +18,14 @@ rendered properly in your Markdown viewer.
1818
[DeLoRA](https://huggingface.co/papers/2503.18225) is a parameter-efficient fine-tuning technique that implicitly maintains a Frobenius boundary with respect to the pretrained weights by normalizing and scaling learnable low-rank matrices. This effectively decouples the learning of directions (BA term) and magnitude (boundary term) of the weight updates, avoiding catastrophic shifts in the adapted weights and enhancing robustness to hyperparameter choices.
1919

2020
Note:
21-
- use 10-100x larger learning rate than standard LoRA variants (typical values from 1e-3/1e-2/..)
22-
- do not set a too small initial boundary parameter lambda (typical values are around 10/15/..)
23-
- setting different lambdas to different layers is possible
21+
- use a learning rate 10-100x larger than for standard LoRA variants (typical values from 1e-3/1e-2/..)
22+
- ensure the initial boundary parameter lambda is not too small (typical values around 10/15/..). Setting different lambdas to different layers is possible
23+
24+
DeLoRA currently has the following constraints:
25+
- Only nn.Linear layers are supported.
26+
- Quantized layers are not supported.
27+
28+
If these constraints don't work for your use case, consider other methods instead.
2429

2530
The abstract from the paper is:
2631

0 commit comments

Comments
 (0)