Grid cells are believed to play an important role in both spatial and
non-spatial cognition tasks. A recent study observed the emergence of grid
cells in an LSTM for path integration. The connection b
Despite the remarkable performance, Deep Neural Networks (DNNs) behave as
black-boxes hindering user trust in Artificial Intelligence (AI) systems.
Research on opening black-box DNN can be broadly cat
Feature propagation in Deep Neural Networks (DNNs) can be associated to
nonlinear discrete dynamical systems. The novelty, in this paper, lies in
letting the discretization parameter (time step-size)
Modern Deep Neural Networks (DNNs) require significant memory to store
weight, activations, and other intermediate tensors during training. Hence,
many models do not fit one GPU device or can be train
We propose an extensive study into the effects of differentially private training (dp) on differentially private deep neural networks (dnns), especially on medicalimaging applications, on the data set of the aptos dataset.
We study the effects of differentially private training on the interpretability of these models, and how the application of differentially private training affects the quality of interpretations.
We propose in this work the gradient-enhanced deep neural networks (DNNs)
approach for function approximations and uncertainty quantification. More
precisely, the proposed approach adopts both the fun
On-line Precision scalability of the deep neural networks(DNNs) is a critical
feature to support accuracy and complexity trade-off during the DNN inference.
In this paper, we propose dual-precision DN
Deep neural networks (DNNs) demonstrate superior performance in various
fields, including scrutiny and security. However, recent studies have shown
that DNNs are vulnerable to backdoor attacks. Severa
Bringing deep neural networks (DNNs) into safety critical applications such
as automated driving, medical imaging and finance, requires a thorough
treatment of the model's uncertainties. Training deep
In this paper, we evaluate the quality of knowledge representations encoded
in deep neural networks (DNNs) for 3D point cloud processing. We propose a
method to disentangle the overall model vulnerabi
Collaborative Filtering (CF) is widely used in recommender systems to model
user-item interactions. With the great success of Deep Neural Networks (DNNs)
in various fields, advanced works recently hav
ADCME is a novel computational framework to solve inverse problems involving
physical simulations and deep neural networks (DNNs). This paper benchmarks its
capability to learn spatially-varying physi
This is the Proceedings of ICML 2021 Workshop on Theoretic Foundation,
Criticism, and Application Trend of Explainable AI. Deep neural networks (DNNs)
have undoubtedly brought great success to a wide
Large pre-trained deep neural networks (dnns) have revolutionized the field of computer vision (cv), but application in industry is often precluded for three reasons : 1) large pre-trained dnns are built on hundreds of millions of parameters, making deployment on many devices impossible, 2) the underlying dataset for pre-training consists of general objects, while industrial cases often consist of very specific objects, such as structures on solar wafers, 3) potentially biased pre-trained dnns raise legal issues for companies.
As a remedy, we study neural networks for cv that we train from scratch.
In this paper, we provide a comprehensive overview on understanding, visualization, and explanation of the internal and overall behavior of deep neural networks (including deep learning architectures).
The aim of this paper is to provide a comprehensive overview on understanding, visualization, and explanation of the internal and overall behavior of deep neural networks (including deep learning architectures).
Deep neural networks (dnns) have been successfully applied to many real-world problems, but a complete understanding of their dynamical and computational principles is still lacking.
Here, by weaving together theories of heavy-tailed random matrices and non-equilibrium statistical physics, we develop a new type of meanfield theory for dnns which predicts that heavy-tailed weights enable the emergence of an extended critical regime without fine-tuning parameters.
Deep neural networks (DNNs) transform stimuli across multiple processing
stages to produce representations that can be used to solve complex tasks, such
as object recognition in images. However, a ful
We propose a 3d heterogeneous manycore architecture for on-chip graph neural network (gnn) training.
The proposed architecture, regraphx, involves heterogeneous reramcrossbars to fulfill the disparate requirements of both deep neural network (dnn) and graphcomputations simultaneously.
Deep Neural Networks (DNNs) are becoming an important tool in modern
computing applications. Accelerating their training is a major challenge and
techniques range from distributed algorithms to low-le
Training deep neural networks (DNNs) efficiently is a challenge due to the
associated highly nonconvex optimization. The backpropagation (backprop)
algorithm has long been the most widely used algorit