Replies: 2 comments 5 replies
-
Padim uses a fixed number of feature embeddings (on Wide ResNet 50 its 550) for calculating the anomalies, while patchcore uses a proportion of all embeddings seen during training (10%). If you have many images, the number of embeddings in patchcore will be significantly larger than in padim, resulting in higher training and espcially validation times |
Beta Was this translation helpful? Give feedback.
-
Thank you for your response ! I now have a better understanding of the models, but I still have a few questions regarding some concepts in PaDiM and PatchCore:
Thank you for your time and for sharing your expertise ! |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi everyone,
I am currently using PatchCore and PaDiM on my own dataset and have a couple of questions about their behavior:
Training Time Differences
When training both algorithms, I noticed that PaDiM takes approximately 8 minutes, whereas PatchCore takes around 2 hours and 30 minutes. Is this difference primarily due to the method used for reducing the memory bank? (For example, PaDiM seems to resize the memory bank randomly, whereas PatchCore uses coreset sub-sampling.)
Validation/Test Behavior with Abnormal Data
As I understand it, training relies only on normal images, while both normal and abnormal images are used for validation and testing.
To explore this, I tested PaDiM with two different datamodules, created using the same training dataset but different abnormal datasets:
o Dataset_1: A square of 5 "dead" pixels (black ones) added to the images.
o Dataset_2: A square of 1 "dead" pixel added to the images.
The model trained with Dataset_1 performs better on its respective validation/test set. However, when I tested this model on Dataset_2, its performance was worse than the model trained with Dataset_2.
This seems counterintuitive since the training is done only on normal data, which is identical for both models. Could this behavior be due to differences in thresholds between the models, resulting in the observed performance discrepancies?
Thank you in advance for your insights!
Beta Was this translation helpful? Give feedback.
All reactions