You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
**pyDVL** collects algorithms for **Data Valuation** and **Influence Function** computation.
@@ -332,7 +318,8 @@ We currently implement the following papers:
332
318
- Schioppa, Andrea, Polina Zablotskaia, David Vilar, and Artem Sokolov.
333
319
[Scaling Up Influence Functions](http://arxiv.org/abs/2112.03052).
334
320
In Proceedings of the AAAI-22. arXiv, 2021.
335
-
321
+
- James Martens, Roger Grosse, [Optimizing Neural Networks with Kronecker-factored Approximate Curvature](https://arxiv.org/abs/1503.05671), International Conference on Machine Learning (ICML), 2015.
322
+
- George, Thomas, César Laurent, Xavier Bouthillier, Nicolas Ballas, Pascal Vincent, [Fast Approximate Natural Gradient Descent in a Kronecker-factored Eigenbasis](https://arxiv.org/abs/1806.03884), Advances in Neural Information Processing Systems 31,2018.
abstract = {Data valuation is a powerful framework for providing statistical insights into which data are beneficial or detrimental to model training. Many Shapley-based data valuation methods have shown promising results in various downstream tasks, however, they are well known to be computationally challenging as it requires training a large number of models. As a result, it has been recognized as infeasible to apply to large datasets. To address this issue, we propose Data-OOB, a new data valuation method for a bagging model that utilizes the out-of-bag estimate. The proposed method is computationally efficient and can scale to millions of data by reusing trained weak learners. Specifically, Data-OOB takes less than $2.25$ hours on a single CPU processor when there are $10^6$ samples to evaluate and the input dimension is $100$. Furthermore, Data-OOB has solid theoretical interpretations in that it identifies the same important data point as the infinitesimal jackknife influence function when two different points are compared. We conduct comprehensive experiments using 12 classification datasets, each with thousands of sample sizes. We demonstrate that the proposed method significantly outperforms existing state-of-the-art data valuation methods in identifying mislabeled data and finding a set of helpful (or harmful) data points, highlighting the potential for applying data values in real-world applications.}
345
+
}
346
+
347
+
@article{george2018fast,
348
+
title={Fast approximate natural gradient descent in a kronecker factored eigenbasis},
349
+
author={George, Thomas and Laurent, C{\'e}sar and Bouthillier, Xavier and Ballas, Nicolas and Vincent, Pascal},
350
+
journal={Advances in Neural Information Processing Systems},
351
+
volume={31},
352
+
year={2018}
353
+
}
354
+
355
+
@inproceedings{martens2015optimizing,
356
+
title={Optimizing neural networks with kronecker-factored approximate curvature},
357
+
author={Martens, James and Grosse, Roger},
358
+
booktitle={International conference on machine learning},
0 commit comments