ResNet or DenseNet? Nowadays, most deep learning based approaches are
implemented with seminal backbone networks, among them the two arguably most
famous ones are ResNet and DenseNet. Despite their co
Deep ResNet architectures have achieved state of the art performance on many
tasks. While they solve the problem of gradient vanishing, they might suffer
from gradient exploding as the depth becomes l
The trend towards increasingly deep neural networks has been driven by a
general observation that increasing depth increases the performance of a
network. Recently, however, evidence has been amassing
We propose a simple yet effective approach that can autonomously learn complex behaviors directly from proprioceptive inputs.
Our approach fuses featuresextracted from pre-trained resnet into the standard reinforcement learning pipeline and delivers results comparable to learning directly from the state.
The ResNet and its variants have achieved remarkable successes in various
computer vision tasks. Despite its success in making gradient flow through
building blocks, the simple shortcut connection mec
Quantization has become a popular technique to compress neural networks and
reduce compute cost, but most prior work focuses on studying quantization
without changing the network size. Many real-world
Melanoma is the most malignant skin tumor and usually cancerates from normal
moles, which is difficult to distinguish benign from malignant in the early
stage. Therefore, many machine learning methods
Regularization plays a major role in modern deep learning. From classic
techniques such as L1,L2 penalties to other noise-based methods such as
Dropout, regularization often yields better generalizati
We present a novel joint geometric-neural networks algorithm for large deformation metric mapping (lddmm) in deformable registration problems of 3d shapes under complex topology-preserving transformations.
The central idea is to represent time-dependent velocity fields as fully connected neural networks (building blocks) and derive optimal weights by minimizing a regularized loss function.
We discuss relations between Residual Networks (ResNet), Recurrent Neural Networks (RNNs) and the primate visual cortex. We begin with the observation that a shallow RNN is exactly equivalent to a ver
We propose the spike-element-wise (sew) residual block (resnet) to realize residual learning in deep spiking neural networks (snns).
We prove that the spike-element-wise (sew) resnet can easily implement identity mapping and overcome the vanishing/exploding gradient problems of previous spiking resnet.
This paper addresses representational block named Hierarchical-Split Block,
which can be taken as a plug-and-play block to upgrade existing convolutional
neural networks, improves model performance si
While image classification models have recently continued to advance, most
downstream applications such as object detection and semantic segmentation
still employ ResNet variants as the backbone netwo
The influential Residual Networks designed by He et al. remain the
gold-standard architecture in numerous scientific publications. They typically
serve as the default architecture in studies, or as ba
In this paper, we aim to understand Residual Network (ResNet) in a
scientifically sound way by providing a bridge between ResNet and Feynman path
integral. In particular, we prove that the effect of r
In our work, we bridge deep neural network design with numerical differential
equations. We show that many effective networks, such as ResNet, PolyNet,
FractalNet and RevNet, can be interpreted as dif
High false alarm rate and low detection rate are the major sticking points for unknown threat perception. To address the problems, in the paper, we present a densely connected residual network (Densel
In recent years, synthetic speech generated by advanced text-to-speech (TTS)
and voice conversion (VC) systems has caused great harms to automatic speaker
verification (ASV) systems, urging us to desi
Overfitting is a crucial problem in deep neural networks, even in the latest
network architectures. In this paper, to relieve the overfitting effect of
ResNet and its improvements (i.e., Wide ResNet,