CEBRA Interpretability #143
Closed
xanderladd
started this conversation in
General
Replies: 1 comment 3 replies
-
Hi @xanderladd , great question! Please have a look at this earlier discussion on this topic, we are actively working on it. Here is our latest work on the topic which we will integrate into CEBRA: https://sslneurips23.github.io/paper_pdfs/paper_80.pdf |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Thanks for this tool and for the support you give on github!
I am able to produce embeddings of neural data that have a clear interpretation in terms of behavior. But I am wondering if there is a way to do something like PCA loadings in order to understand the contribution of an individual neuron to the embeddings. I know this is a tall order because the weights of the neural network are not intuitively interpretable. Is this worth thinking about? Or should I just consider the positive / negative sample neural nets to be black boxes?
Some ideas I had for this were:
Ultimately, what would be great to know is some way of quantifying a single neuron (or a subset of neurons) contribution to the embeddings. I don't expect to make strong causal arguments this way, but correlational evidence for further hypothesis discovery would be good.
Beta Was this translation helpful? Give feedback.
All reactions