You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+41-1Lines changed: 41 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -132,7 +132,47 @@ Example: Train.jsonl
132
132
133
133
## Supported Models
134
134
135
-
Current supported and tested models are `Llama3` (8B configuration has been tested) and `GPTBigCode`.
135
+
- For each tuning technique, we run testing on a single large model of each architecture type and claim support for the smaller models. For example, with QLoRA technique, we tested on granite-34b GPTBigCode and claim support for granite-20b-multilingual.
136
+
137
+
- LoRA Layers supported : All the linear layers of a model + output `lm_head` layer. Users can specify layers as a list or use `all-linear` as a shortcut. Layers are specific to a model architecture and can be specified as noted [here](https://github.com/foundation-model-stack/fms-hf-tuning?tab=readme-ov-file#lora-tuning-example)
138
+
139
+
- Legend:
140
+
141
+
✅ Ready and available
142
+
143
+
✔️ Ready and available - compatible architecture (*see first bullet point above)
144
+
145
+
🚫 Not supported
146
+
147
+
? May be supported, but not tested
148
+
149
+
Model Name & Size | Model Architecture | Full Finetuning | Low Rank Adaptation (i.e. LoRA) | qLoRA(quantized LoRA) |
0 commit comments