-
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 1 reply
-
|
Beta Was this translation helpful? Give feedback.
-
But the implementation is simple: we sample Z according to the feature distribution of the training set for every feature dimension (I only experimented on categorical features, in which case we just sample by class probability of the categorical feature); and then apply the mask F using that formula. So for each feature dimension, when F=0, you get the randomly sampled Z according to the marginal, which means the feature is not important. When F=1, you recover X which means the feature dimension is important. And then we can threshold out the unimportant dimensions just as how we threshold out unimportant edges.
|
Beta Was this translation helpful? Give feedback.
But the implementation is simple: we sample Z according to the feature distribution of the training set for every feature dimension (I only experimented on categorical features, in which case we just sample by class probability of the categorical feature); and then apply the mask F using that formula. So for each feature dimension, when F=0, you get the randomly sampled Z according to the marginal, which means the feature is not important. When F=1, you rec…