Abstract: The rise in deep neural networks (DNNs) has led to increased interest in explaining their predictions. While many methods for this exist, there is currently no consensus on how to evaluate ...
SVGD is a general purpose variational inference algorithm that forms a natural counterpart of gradient descent for optimization. SVGD iteratively transports a set of particles to match with the target ...
Training very deep neural networks requires a lot of memory. Using the tools in this package, developed jointly by Tim Salimans and Yaroslav Bulatov, you can trade off some of this memory usage with ...
Numerical optimization is an important tool in the field of computational physics in general and in nano-optics in specific. It has attracted attention with the increase in complexity of structures ...
Experimentally, we generated this thermal gradient by irradiating the sample with a pump beam resonant with the optical absorption of efficient light to heat transducers. In our case, we used ...