In this paper we propose a lightning fast graph embedding method called graph
encoder embedding. The proposed method has a linear computational complexity
and the capacity to process billions of edges
We present models for encoding sentences into embedding vectors that
specifically target transfer learning to other NLP tasks. The models are
efficient and result in accurate performance on diverse tr
Spiking neural networks are a promising approach towards next-generation
models of the brain in computational neuroscience. Moreover, compared to
classic artificial neural networks, they could serve a
We consider the problem of information compression from high dimensional
data. Where many studies consider the problem of compression by non-invertible
transformations, we emphasize the importance of
State-of-the-art encoder-decoder models (e.g. for machine translation (MT) or
speech recognition (ASR)) are constructed and trained end-to-end as an atomic
unit. No component of the model can be (re-)
Variational Auto-Encoders (VAEs) have become very popular techniques to
perform inference and learning in latent variable models as they allow us to
leverage the rich representational power of neural
Quantum mechanical unitarity in our universe is challenged both by the notion
of the big bang, in which nothing transforms into something, and the expansion
of space, in which something transforms int
We introduce the idea of using the reach of the manifold spanned by the decoder to determine if an optimal encoder exists for a given dataset and decoder.
We develop a local generalization of this reach and propose a numerical estimator thereof.
We propose a compact pretrained deep neural network, transformer encoder for social science (tess), explicitly designed to tackle text processing tasks in social science research.
Using two validation tests, we demonstrate that the superiority of our model over other pretrained deep neural network models on social science text processing tasks is 16.7% on average when the number of training samples is limited (1,000 training instances).
Efficient k-nearest neighbor search is a fundamental task, foundational for
many problems in NLP. When the similarity is measured by dot-product between
dual-encoder vectors or $\ell_2$-distance, ther
Guetzli is a new JPEG encoder that aims to produce visually indistinguishable
images at a lower bit-rate than other common JPEG encoders. It optimizes both
the JPEG global quantization tables and the
We present a preliminary study investigating rank-one editing as a direct intervention method for behavior deletion requests in encoder-decoder-transformer models.
We propose four editing tasks for neural machine translation and show that the proposed editing algorithm achieves high efficacy, while requiring only a single instance of positive example to fix an erroneous (negative) model behavior.
Foundation models or pre-trained models have substantially improved the
performance of various language, vision, and vision-language understanding
tasks. However, existing foundation models can only p
Mechanical devices such as engines, vehicles, aircrafts, etc., are typically
instrumented with numerous sensors to capture the behavior and health of the
machine. However, there are often external fac
This paper proposes a context-free hierarchical motion encoder-decoder network (hmnet) for vehicletrajectory prediction.
Hmnet first infers the hierarchical difference on motions to encode physically compliant patterns with high expressivity of moving trends and driving intentions.
Based on the idea of introducing intimate information interaction between the two information sources, we propose our siamese attribute-missinggraph auto-encoder (saga).
We entangle the learning of attribute and structure embedding by introducing a siamese network structure to share the parameters learned by both processes, which allows the network training to benefit from more abundant and diverse information.
The output structure of database-like tables, consisting of values structured
in horizontal rows and vertical columns identifiable by name, can cover a wide
range of NLP tasks. Following this constata
Among the wide variety of image generative models, two models stand out:
Variational Auto Encoders (VAE) and Generative Adversarial Networks (GAN). GANs
can produce realistic images, but they suffer f
In this article, we investigate how probability function gradient ascent over datapoints can be used to process data in order to achieve better clustering.
We propose a simple yet effective method for investigating suitable number of clusters for data, based on the dbscan clustering algorithm, and demonstrate that cluster number determination is facilitated with gradientprocessing.