Skip to content

Commit 371ef79

Browse files
Merge branch 'develop' into feature/simplify-parallel-backend-config
2 parents 65313e7 + 03fd179 commit 371ef79

File tree

9 files changed

+62
-18
lines changed

9 files changed

+62
-18
lines changed

docs/assets/pydvl.bib

Lines changed: 28 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,8 @@ @article{agarwal_secondorder_2017
1010
pages = {1--40},
1111
url = {https://www.jmlr.org/papers/v18/16-491.html},
1212
abstract = {First-order stochastic methods are the state-of-the-art in large-scale machine learning optimization owing to efficient per-iteration complexity. Second-order methods, while able to provide faster convergence, have been much less explored due to the high cost of computing the second-order information. In this paper we develop second-order stochastic methods for optimization problems in machine learning that match the per-iteration cost of gradient based methods, and in certain settings improve upon the overall running time over popular first-order methods. Furthermore, our algorithm has the desirable property of being implementable in time linear in the sparsity of the input data.},
13-
langid = {english}
13+
langid = {english},
14+
keywords = {notion}
1415
}
1516

1617
@inproceedings{bae_if_2022,
@@ -340,6 +341,19 @@ @inproceedings{li_achieving_2022
340341
langid = {english}
341342
}
342343

344+
@article{maleki_bounding_2014,
345+
title = {Bounding the {{Estimation Error}} of {{Sampling-based Shapley Value Approximation}}},
346+
author = {Maleki, Sasan and Tran-Thanh, Long and Hines, Greg and Rahwan, Talal and Rogers, Alex},
347+
date = {2014-02-12},
348+
journaltitle = {ArXiv13064265 Cs},
349+
eprint = {1306.4265},
350+
eprinttype = {arxiv},
351+
eprintclass = {cs},
352+
url = {https://arxiv.org/abs/1306.4265},
353+
urldate = {2020-11-16},
354+
abstract = {The Shapley value is arguably the most central normative solution concept in cooperative game theory. It specifies a unique way in which the reward from cooperation can be "fairly" divided among players. While it has a wide range of real world applications, its use is in many cases hampered by the hardness of its computation. A number of researchers have tackled this problem by (i) focusing on classes of games where the Shapley value can be computed efficiently, or (ii) proposing representation formalisms that facilitate such efficient computation, or (iii) approximating the Shapley value in certain classes of games. For the classical \textbackslash textit\{characteristic function\} representation, the only attempt to approximate the Shapley value for the general class of games is due to Castro \textbackslash textit\{et al.\} \textbackslash cite\{castro\}. While this algorithm provides a bound on the approximation error, this bound is \textbackslash textit\{asymptotic\}, meaning that it only holds when the number of samples increases to infinity. On the other hand, when a finite number of samples is drawn, an unquantifiable error is introduced, meaning that the bound no longer holds. With this in mind, we provide non-asymptotic bounds on the estimation error for two cases: where (i) the \textbackslash textit\{variance\}, and (ii) the \textbackslash textit\{range\}, of the players' marginal contributions is known. Furthermore, for the second case, we show that when the range is significantly large relative to the Shapley value, the bound can be improved (from \$O(\textbackslash frac\{r\}\{m\})\$ to \$O(\textbackslash sqrt\{\textbackslash frac\{r\}\{m\}\})\$). Finally, we propose, and demonstrate the effectiveness of using stratified sampling for improving the bounds further.}
355+
}
356+
343357
@inproceedings{martens_optimizing_2015,
344358
title = {Optimizing {{Neural Networks}} with {{Kronecker-factored Approximate Curvature}}},
345359
booktitle = {Proceedings of the 32nd {{International Conference}} on {{Machine Learning}}},
@@ -388,19 +402,23 @@ @inproceedings{okhrati_multilinear_2021
388402
keywords = {notion}
389403
}
390404

391-
@inproceedings{schioppa_scaling_2021,
405+
@article{schioppa_scaling_2022,
392406
title = {Scaling {{Up Influence Functions}}},
393407
author = {Schioppa, Andrea and Zablotskaia, Polina and Vilar, David and Sokolov, Artem},
394-
date = {2021-12-06},
408+
date = {2022-06-28},
409+
journaltitle = {Proceedings of the AAAI Conference on Artificial Intelligence},
410+
shortjournal = {Proc. AAAI Conf. Artif. Intell.},
411+
volume = {36},
412+
number = {8},
395413
eprint = {2112.03052},
396414
eprinttype = {arxiv},
397-
eprintclass = {cs},
398-
publisher = {arXiv},
399-
doi = {10.48550/arXiv.2112.03052},
400-
url = {https://arxiv.org/abs/2112.03052},
401-
urldate = {2023-03-10},
402-
abstract = {We address efficient calculation of influence functions for tracking predictions back to the training data. We propose and analyze a new approach to speeding up the inverse Hessian calculation based on Arnoldi iteration. With this improvement, we achieve, to the best of our knowledge, the first successful implementation of influence functions that scales to full-size (language and vision) Transformer models with several hundreds of millions of parameters. We evaluate our approach on image classification and sequence-to-sequence tasks with tens to a hundred of millions of training examples. Our code will be available at https://github.com/google-research/jax-influence.},
403-
eventtitle = {{{AAAI-22}}},
415+
pages = {8179--8186},
416+
issn = {2374-3468},
417+
doi = {10.1609/aaai.v36i8.20791},
418+
url = {https://ojs.aaai.org/index.php/AAAI/article/view/20791},
419+
urldate = {2024-03-30},
420+
abstract = {We address efficient calculation of influence functions for tracking predictions back to the training data. We propose and analyze a new approach to speeding up the inverse Hessian calculation based on Arnoldi iteration. With this improvement, we achieve, to the best of our knowledge, the first successful implementation of influence functions that scales to full-size (language and vision) Transformer models with several hundreds of millions of parameters. We evaluate our approach in image classification and sequence-to-sequence tasks with tens to a hundred of millions of training examples. Our code is available at https://github.com/google-research/jax-influence.},
421+
langid = {english},
404422
keywords = {notion}
405423
}
406424

docs/getting-started/glossary.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ Terms in data valuation and influence functions:
1414
The Arnoldi method approximately computes eigenvalue, eigenvector pairs of
1515
a symmetric matrix. For influence functions, it is used to approximate
1616
the [iHVP][inverse-hessian-vector-product].
17-
Introduced by [@schioppa_scaling_2021] in the context of influence functions.
17+
Introduced by [@schioppa_scaling_2022] in the context of influence functions.
1818

1919
* [Implementation (torch)
2020
][pydvl.influence.torch.influence_function_model.ArnoldiInfluence]

docs/getting-started/methods.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -56,7 +56,7 @@ We currently implement the following methods:
5656
[@agarwal_secondorder_2017].
5757

5858
- [**Arnoldi Influence**][pydvl.influence.torch.ArnoldiInfluence]
59-
[@schioppa_scaling_2021].
59+
[@schioppa_scaling_2022].
6060

6161
- [**EKFAC Influence**][pydvl.influence.torch.EkfacInfluence]
6262
[@george_fast_2018;@martens_optimizing_2015].

docs/influence/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ intractable for large models like deep neural networks. Much of the recent
2424
research tackles this issue using approximations, like a Neuman series
2525
[@agarwal_secondorder_2017], with the most successful solution using a low-rank
2626
approximation that iteratively finds increasing eigenspaces of the Hessian
27-
[@schioppa_scaling_2021].
27+
[@schioppa_scaling_2022].
2828

2929
pyDVL implements several methods for the efficient computation of the IF for
3030
machine learning. In the examples we document some of the difficulties that can

docs/influence/influence_function_model.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -107,7 +107,7 @@ approximated by
107107

108108
where \(D\) is a diagonal matrix with the top (in absolute value) eigenvalues of
109109
the Hessian and \(V\) contains the corresponding eigenvectors. See also
110-
[@schioppa_scaling_2021].
110+
[@schioppa_scaling_2022].
111111

112112
```python
113113
from pydvl.influence.torch import ArnoldiInfluence
Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,18 @@
1+
{% set site_domain = config.site_url | replace('https://', '') | replace('http://', '') | trim('/') %}
2+
{% set dev_domain = config.dev_addr | replace('https://', '') | replace('http://', '') | trim('/') %}
3+
{% set develop = site_domain.startswith(dev_domain) %}
4+
{% if not develop %}
5+
{% set hostname = config.extra.analytics.hostname | default(site_domain) %}
6+
{% set script_domain = config.extra.analytics.script_domain | default('scripts.simpleanalyticscdn.com') %}
7+
{% set noscript_domain = config.extra.analytics.noscript_domain | default('queue.simpleanalyticscdn.com') %}
8+
{% set collect_dnt = config.extra.analytics.collect_dnt | default('false') %}
9+
<script async defer
10+
src="https://{{ script_domain }}/latest.js"
11+
data-collect-dnt="true"
12+
data-hostname="{{ hostname }}"></script>
13+
<noscript>
14+
<img src="https://{{ noscript_domain }}/noscript.gif?collect-dnt={{ collect_dnt }}&hostname={{ hostname }}&path={{ page.url }}"
15+
alt=""
16+
referrerpolicy="no-referrer-when-downgrade">
17+
</noscript>
18+
{% endif %}

mkdocs.yml

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -214,6 +214,12 @@ extra:
214214
link: https://twitter.com/aai_transferlab
215215
- icon: fontawesome/brands/linkedin
216216
link: https://de.linkedin.com/company/appliedai-institute-for-europe-ggmbh
217+
analytics:
218+
provider: simpleanalytics
219+
hostname: pydvl.org
220+
script_domain: scripts.simpleanalyticscdn.com
221+
noscript_domain: queue.simpleanalyticscdn.com
222+
collect_dnt: true
217223

218224
markdown_extensions:
219225
- abbr

requirements-docs.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ mkdocs-gen-files
1111
mkdocs-git-revision-date-localized-plugin
1212
mkdocs-glightbox==0.3.7
1313
mkdocs-literate-nav
14-
mkdocs-material[imaging]==9.5.14
14+
mkdocs-material[imaging]==9.5.16
1515
mkdocs-section-index==0.3.8
1616
mkdocs-macros-plugin
1717
mknotebooks==0.8.0

requirements-notebooks.txt

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,7 @@
1-
torch==2.0.1
2-
torchvision==0.15.2
31
datasets==2.14.6
2+
distributed==2024.4.0
43
pillow==10.0.1
5-
transformers==4.35.0
4+
torch==2.0.1
5+
torchvision==0.15.2
6+
transformers==4.36.0
7+
zarr==2.17.1

0 commit comments

Comments
 (0)