The practical success of overparameterized neural networks has motivated the
recent scientific study of interpolating methods, which perfectly fit their
training data. Certain interpolating methods, i
We introduce a method to train Binarized Neural Networks (BNNs) - neural
networks with binary weights and activations at run-time. At training-time the
binary weights and activations are used for comp
In this paper, we propose a method for uncertainty estimation in neural networks called variational neural network (vnn) that, instead of considering a distribution over weights, generates parameters for the output distribution of a layer by transforming its inputs with learnable sub-layers.
In uncertainty quality estimation experiments, we show that vnns achieve better uncertainty quality than monte carlo dropout or bayesby backpropagation methods.
Boosting is a method for learning an ensemble of classifiers by linearly combining many ``weak''hypotheses, each of which may be only moderately accurate.
While boosting has been shown to be very effective for decision trees, its impact on neural networks has not been extensively studied.
We empirically evaluate common assumptions about neural networks that are
widely held by practitioners and theorists alike. In this work, we: (1) prove
the widespread existence of suboptimal local min
Hyperbolic spaces have recently gained momentum in the context of machine
learning due to their high capacity and tree-likeliness properties. However,
the representational power of hyperbolic geometry
Wide neural networks with random weights and biases are Gaussian processes,
as originally observed by Neal (1995) and more recently by Lee et al. (2018)
and Matthews et al. (2018) for deep fully-conne
Accurate models of the world are built upon notions of its underlying
symmetries. In physics, these symmetries correspond to conservation laws, such
as for energy and momentum. Yet even though neural
Based on the assumption that there exists a neural network that efficiently
represents a set of Boolean functions between all binary inputs and outputs, we
propose a process for developing and deployi
Graph Neural Networks (GNNs) are the first choice for learning algorithms on
graph data. GNNs promise to integrate (i) node features as well as (ii) edge
information in an end-to-end learning algorith
We describe a novel family of models of multi-layer feedforward neuralnetworks in which the activation functions are encoded via penalties in the training problem.
Our approach is based on representing a non-decreasing activation function as the argmin of an appropriate convex optimiza-tion problem.
This work connects models for virus spread on networks with their equivalent
neural network representations. Based on this connection, we propose a new
neural network architecture, called Transmission
In recent times, neural networks have become a powerful tool for the analysis
of complex and abstract data models. However, their introduction intrinsically
increases our uncertainty about which featu
Deep neural networks are becoming increasingly popular in approximating
arbitrary functions from noisy data. But wider adoption is being hindered by
the need to explain such models and to impose addit
We study how neural networks trained by gradient descent extrapolate, i.e.,
what they learn outside the support of training distribution. Previous works
report mixed empirical results when extrapolati
This PhD thesis combines two of the most exciting research areas of the last
decades: quantum computing and machine learning. We introduce dissipative
quantum neural networks (DQNNs), which are design
Spiking neural networks are a promising approach towards next-generation
models of the brain in computational neuroscience. Moreover, compared to
classic artificial neural networks, they could serve a