|
1 | 1 | @misc{blondelElementsDifferentiableProgramming2024, |
2 | | - title = {The {{Elements}} of {{Differentiable Programming}}}, |
3 | | - author = {Blondel, Mathieu and Roulet, Vincent}, |
4 | | - year = {2024}, |
5 | | - month = mar, |
6 | | - number = {arXiv:2403.14606}, |
7 | | - eprint = {2403.14606}, |
8 | | - primaryclass = {cs}, |
9 | | - publisher = {arXiv}, |
10 | | - doi = {10.48550/arXiv.2403.14606}, |
11 | | - url = {http://arxiv.org/abs/2403.14606}, |
12 | | - urldate = {2024-03-22}, |
13 | | - abstract = {Artificial intelligence has recently experienced remarkable advances, fueled by large models, vast datasets, accelerated hardware, and, last but not least, the transformative power of differentiable programming. This new programming paradigm enables end-to-end differentiation of complex computer programs (including those with control flows and data structures), making gradient-based optimization of program parameters possible. As an emerging paradigm, differentiable programming builds upon several areas of computer science and applied mathematics, including automatic differentiation, graphical models, optimization and statistics. This book presents a comprehensive review of the fundamental concepts useful for differentiable programming. We adopt two main perspectives, that of optimization and that of probability, with clear analogies between the two. Differentiable programming is not merely the differentiation of programs, but also the thoughtful design of programs intended for differentiation. By making programs differentiable, we inherently introduce probability distributions over their execution, providing a means to quantify the uncertainty associated with program outputs.}, |
14 | | - archiveprefix = {arXiv}, |
| 2 | + title = {The {{Elements}} of {{Differentiable Programming}}}, |
| 3 | + author = {Blondel, Mathieu and Roulet, Vincent}, |
| 4 | + year = {2024}, |
| 5 | + month = mar, |
| 6 | + number = {arXiv:2403.14606}, |
| 7 | + eprint = {2403.14606}, |
| 8 | + primaryclass = {cs}, |
| 9 | + publisher = {arXiv}, |
| 10 | + doi = {10.48550/arXiv.2403.14606}, |
| 11 | + url = {http://arxiv.org/abs/2403.14606}, |
| 12 | + urldate = {2024-03-22}, |
| 13 | + abstract = {Artificial intelligence has recently experienced remarkable advances, fueled by large models, vast datasets, accelerated hardware, and, last but not least, the transformative power of differentiable programming. This new programming paradigm enables end-to-end differentiation of complex computer programs (including those with control flows and data structures), making gradient-based optimization of program parameters possible. As an emerging paradigm, differentiable programming builds upon several areas of computer science and applied mathematics, including automatic differentiation, graphical models, optimization and statistics. This book presents a comprehensive review of the fundamental concepts useful for differentiable programming. We adopt two main perspectives, that of optimization and that of probability, with clear analogies between the two. Differentiable programming is not merely the differentiation of programs, but also the thoughtful design of programs intended for differentiation. By making programs differentiable, we inherently introduce probability distributions over their execution, providing a means to quantify the uncertainty associated with program outputs.}, |
| 14 | + archiveprefix = {arXiv} |
15 | 15 | } |
16 | 16 | % == BibTeX quality report for blondelElementsDifferentiableProgramming2024: |
17 | 17 | % ? Title looks like it was stored in title-case in Zotero |
18 | 18 |
|
19 | 19 | @article{koolBuyREINFORCESamples2022, |
20 | | - title = {Buy 4 {{REINFORCE Samples}}, {{Get}} a {{Baseline}} for {{Free}}!}, |
21 | | - author = {Kool, Wouter and van Hoof, Herke and Welling, Max}, |
22 | | - year = {2022}, |
23 | | - month = jul, |
24 | | - url = {https://openreview.net/forum?id=r1lgTGL5DE}, |
25 | | - urldate = {2023-04-17}, |
| 20 | + title = {Buy 4 {{REINFORCE Samples}}, {{Get}} a {{Baseline}} for {{Free}}!}, |
| 21 | + author = {Kool, Wouter and van Hoof, Herke and Welling, Max}, |
| 22 | + year = {2022}, |
| 23 | + month = jul, |
| 24 | + journal = {ICLR}, |
| 25 | + url = {https://openreview.net/forum?id=r1lgTGL5DE}, |
| 26 | + urldate = {2023-04-17}, |
26 | 27 | abstract = {REINFORCE can be used to train models in structured prediction settings to directly optimize the test-time objective. However, the common case of sampling one prediction per datapoint (input) is data-inefficient. We show that by drawing multiple samples (predictions) per datapoint, we can learn with significantly less data, as we freely obtain a REINFORCE baseline to reduce variance. Additionally we derive a REINFORCE estimator with baseline, based on sampling without replacement. Combined with a recent technique to sample sequences without replacement using Stochastic Beam Search, this improves the training procedure for a sequence model that predicts the solution to the Travelling Salesman Problem.}, |
27 | | - langid = {english}, |
28 | | - language = {en}, |
| 28 | + langid = {english}, |
| 29 | + language = {en} |
29 | 30 | } |
30 | 31 | % == BibTeX quality report for koolBuyREINFORCESamples2022: |
31 | 32 | % Missing required field 'journal' |
32 | 33 | % ? Title looks like it was stored in title-case in Zotero |
33 | 34 | % ? unused Library catalog ("openreview.net") |
34 | 35 |
|
35 | 36 | @article{mohamedMonteCarloGradient2020, |
36 | | - title = {Monte {{Carlo Gradient Estimation}} in {{Machine Learning}}}, |
37 | | - author = {Mohamed, Shakir and Rosca, Mihaela and Figurnov, Michael and Mnih, Andriy}, |
38 | | - year = {2020}, |
39 | | - journal = {Journal of Machine Learning Research}, |
40 | | - volume = {21}, |
41 | | - number = {132}, |
42 | | - pages = {1--62}, |
43 | | - issn = {1533-7928}, |
44 | | - url = {http://jmlr.org/papers/v21/19-346.html}, |
45 | | - urldate = {2022-10-21}, |
46 | | - abstract = {This paper is a broad and accessible survey of the methods we have at our disposal for Monte Carlo gradient estimation in machine learning and across the statistical sciences: the problem of computing the gradient of an expectation of a function with respect to parameters defining the distribution that is integrated; the problem of sensitivity analysis. In machine learning research, this gradient problem lies at the core of many learning problems, in supervised, unsupervised and reinforcement learning. We will generally seek to rewrite such gradients in a form that allows for Monte Carlo estimation, allowing them to be easily and efficiently used and analysed. We explore three strategies---the pathwise, score function, and measure-valued gradient estimators---exploring their historical development, derivation, and underlying assumptions. We describe their use in other fields, show how they are related and can be combined, and expand on their possible generalisations. Wherever Monte Carlo gradient estimators have been derived and deployed in the past, important advances have followed. A deeper and more widely-held understanding of this problem will lead to further advances, and it is these advances that we wish to support.}, |
| 37 | + title = {Monte {{Carlo Gradient Estimation}} in {{Machine Learning}}}, |
| 38 | + author = {Mohamed, Shakir and Rosca, Mihaela and Figurnov, Michael and Mnih, Andriy}, |
| 39 | + year = {2020}, |
| 40 | + journal = {Journal of Machine Learning Research}, |
| 41 | + volume = {21}, |
| 42 | + number = {132}, |
| 43 | + pages = {1--62}, |
| 44 | + issn = {1533-7928}, |
| 45 | + url = {http://jmlr.org/papers/v21/19-346.html}, |
| 46 | + urldate = {2022-10-21}, |
| 47 | + abstract = {This paper is a broad and accessible survey of the methods we have at our disposal for Monte Carlo gradient estimation in machine learning and across the statistical sciences: the problem of computing the gradient of an expectation of a function with respect to parameters defining the distribution that is integrated; the problem of sensitivity analysis. In machine learning research, this gradient problem lies at the core of many learning problems, in supervised, unsupervised and reinforcement learning. We will generally seek to rewrite such gradients in a form that allows for Monte Carlo estimation, allowing them to be easily and efficiently used and analysed. We explore three strategies---the pathwise, score function, and measure-valued gradient estimators---exploring their historical development, derivation, and underlying assumptions. We describe their use in other fields, show how they are related and can be combined, and expand on their possible generalisations. Wherever Monte Carlo gradient estimators have been derived and deployed in the past, important advances have followed. A deeper and more widely-held understanding of this problem will lead to further advances, and it is these advances that we wish to support.} |
47 | 48 | } |
48 | 49 | % == BibTeX quality report for mohamedMonteCarloGradient2020: |
49 | 50 | % ? Title looks like it was stored in title-case in Zotero |
|
0 commit comments