-
Notifications
You must be signed in to change notification settings - Fork 5
General
Joseph cheng edited this page May 11, 2018
·
28 revisions
- Net2Net
- Can Active Memory Replace Attention?
- Extensions and Limitations of the Neural GPU
- Evolution Strategies
- Deep Forest: Towards An Alternative to Deep Neural Networks
- A Regularized Framework for Sparse and Structured Neural Attention
- Self-Normalizing Neural Networks
- FreezeOut: Accelerate Training by Progressively Freezing Layers
- Super-Convergence: Very Fast Training of Residual Networks Using Large Learning Rates
- Memory-based Parameter Adaptation
- Meta-Learning via Feature-Label Memory Network
- Adversarial Contrastive Estimation
- Variational Dropout and the Local Reparameterization Trick
- Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning
- Concrete Dropout
- Improving Variational Auto-Encodersusing Householder Flow
- Multiplicative Normalizing Flows for Variational Bayesian Neural Networks
- Variational Attention for Sequence-to-Sequence Models