We propose a novel hyperbolic graph convolutional network named lorentzian graph convolutional network (lgcn), which rigorously guarantees the learned node features follow the hyperbolic geometry.
Specifically, we rebuild the graph operations of hyperbolic graph convolutional networks with the lorentzian version, e.g., the feature transformation and non-linear activation.
Graph neural networks (GNNs) have significantly improved the representation
power for graph-structured data. Despite of the recent success of GNNs, the
graph convolution in most GNNs have two limitati
The aim of this work is to develop a fully-distributed algorithmic framework
for training graph convolutional networks (GCNs). The proposed method is able
to exploit the meaningful relational structur
Graph Convolutional Networks (GCNs) and their variants have experienced
significant attention and have become the de facto methods for learning graph
representations. GCNs derive inspiration primarily
Graph Convolutional Networks (GCNs) achieve an impressive performance due to
the remarkable representation ability in learning the graph information.
However, GCNs, when implemented on a deep network,
We develop new geodesic tools thatallow for extending neural network operations into geodesically disconnected semi-riemannian manifolds.
Our methodprovides a geometric inductive bias that is sufficiently flexible to model mixed heterogeneous topologies like hierarchical graphs with cycles.
We present a generalization of graph convolutional networks by generalizing
the diffusion operation underlying this class of graph neural networks. These
sheaf neural networks are based on the sheaf L
Spectral graph convolutional neural networks (CNNs) require approximation to
the convolution to alleviate the computational complexity, resulting in
performance loss. This paper proposes the topology
Text classification is an important and classical problem in natural language
processing. There have been a number of studies that applied convolutional
neural networks (convolution on regular grid, e
Graph deep learning has recently emerged as a powerful ML concept allowing to generalize successful deep neural architectures to non-Euclidean structured data. Such methods have shown promising result
Graph convolution operator of the GCN model is originally motivated from a
localized first-order approximation of spectral graph convolutions. This work
stands on a different view; establishing a \tex
Convolutional neural networks (CNNs) have been attracting increasing attention in hyperspectral (HS) image classification, owing to their ability to capture spatial-spectral feature representations. N
Knowledge graphs enable a wide variety of applications, including question
answering and information retrieval. Despite the great effort invested in their
creation and maintenance, even the largest (e
Graph convolutional networks (GCNs) are a powerful deep learning approach for
graph-structured data. Recently, GCNs and subsequent variants have shown
superior performance in various application areas
Graph Convolutional Networks (GCN) with multi-hop aggregation is more
expressive than one-hop GCN but suffers from higher model complexity. Finding
the shortest aggregation range that achieves compara
Potential-based reward shaping provides an approach for designing good reward
functions, with the purpose of speeding up learning. However, automatically
finding potential functions for complex enviro
Graph convolutional networks produce good predictions of unlabeled samples
due to its transductive label propagation. Since samples have different
predicted confidences, we take high-confidence predic
We present GraphMix, a regularization method for Graph Neural Network based
semi-supervised object classification, whereby we propose to train a
fully-connected network jointly with the graph neural n
Recent advances in graph convolutional networks have significantly improved
the performance of chemical predictions, raising a new research question: "how
do we explain the predictions of graph convol
Graph Convolutional Network (GCN) has exhibited strong empirical performance
in many real-world applications. The vast majority of existing works on GCN
primarily focus on the accuracy while ignoring