-
Notifications
You must be signed in to change notification settings - Fork 558
Description
Hi team,
Thank you very much for the excellent work on TabPFN! It has been extremely helpful in my research.
I have a question regarding fine-tuning TabPFN for regression with high-dimensional pretrained embeddings.
I am working with protein-related data where each sample is represented by embeddings from pretrained models, typically 1000–2000 features per sample.
Using standard TabPFN in-context inference works well and is memory-efficient.
However, when I try to use FinetunedTabPFNRegressor, GPU memory usage explodes during training (e.g. in the KV projection step such as
torch.einsum("... s, j h d s -> ... j h d", x_kv, w_kv)), exceeding 48 GB VRAM (RTX Ada 6000).
My questions are:
Are there recommended best practices for fine-tuning TabPFN with high-dimensional pretrained embeddings?
Is dimensionality reduction generally expected before fine-tuning, and are there suggested target dimensions?
Or is fine-tuning simply not recommended for such high-dimensional inputs, with in-context inference being the intended use case?
Any guidance would be greatly appreciated.
Thanks again for your great work!