Skip to content

Finetune regression with high-dimensional features #716

@ivandon15

Description

@ivandon15

Hi team,

Thank you very much for the excellent work on TabPFN! It has been extremely helpful in my research.

I have a question regarding fine-tuning TabPFN for regression with high-dimensional pretrained embeddings.
I am working with protein-related data where each sample is represented by embeddings from pretrained models, typically 1000–2000 features per sample.

Using standard TabPFN in-context inference works well and is memory-efficient.
However, when I try to use FinetunedTabPFNRegressor, GPU memory usage explodes during training (e.g. in the KV projection step such as
torch.einsum("... s, j h d s -> ... j h d", x_kv, w_kv)), exceeding 48 GB VRAM (RTX Ada 6000).

My questions are:
Are there recommended best practices for fine-tuning TabPFN with high-dimensional pretrained embeddings?
Is dimensionality reduction generally expected before fine-tuning, and are there suggested target dimensions?
Or is fine-tuning simply not recommended for such high-dimensional inputs, with in-context inference being the intended use case?

Any guidance would be greatly appreciated.
Thanks again for your great work!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions