Self-supervised learning has attracted many researchers for its soaring performance on representation learning in the last several years.
In this survey, we take a look into new self-supervised learning methods for representation in computer vision, natural language processing, and graph learning.
Self-supervised Learning (SSL) including the mainstream contrastive learning
has achieved great success in learning visual representations without data
annotations. However, most methods mainly focus
Kidney segmentation is leveraged to define an effective proxy task for kidney segmentation via self-supervised learning.
Evaluation results on a publicly available dataset containing computed tomography (ct) scans of the abdominal region shows that a boost in performance and fast convergence can be had relative to a network trained conventionally from scratch.
Self-training is an effective approach to semi-supervised learning. The key
idea is to let the learner itself iteratively generate "pseudo-supervision" for
unlabeled instances based on its current hyp
Recently, unsupervised adversarial training (AT) has been extensively studied
to attain robustness with the models trained upon unlabeled data. To this end,
previous studies have applied existing supe
In this book chapter, we discuss 10 basic security and privacy problems for the pre-trained encoders in self-supervised learning, including six confidentiality problems, three integrity problems, and one availability problem.
We hope our bookchapter will inspire future research on the security and privacy of self-supervised learning.
The popularisation of neural networks has seen incredible advances in pattern
recognition, driven by the supervised learning of human annotations. However,
this approach is unsustainable in relation t
Ensemble methods have proven to be a powerful technique for boosting modelperformance, uncertainty estimation, and robustness in supervised learning.
Recent advances in self-supervised learning (ssl) enable leveraging large unlabeled corpora for state-of-the-art few-shot and supervised learning performance.
Our work reveals a structured shortcoming of the existing mainstream
self-supervised learning methods. Whereas self-supervised learning frameworks
usually take the prevailing perfect instance level in
We introduce a simple mean-shift algorithm that learns representations by grouping images together withoutcontrasting between them or adopting much of prior on the structure of the image clusters.
Our model achieves 72.4% on imagenet linear evaluation with resnet50 at 200epochs outperforming byol.
Noisy labels, resulting from mistakes in manual labeling or webly data
collecting for supervised learning, can cause neural networks to overfit the
misleading information and degrade the generalizatio
This paper proposes a novel self-supervised learning method for learning
better representations with small batch sizes. Many self-supervised learning
methods based on certain forms of the siamese netw
Self-supervised learning methods are gaining increasing traction in computer
vision due to their recent success in reducing the gap with supervised
learning. In natural language processing (NLP) self-
Graph self-supervised learning has gained increasing attention due to its
capacity to learn expressive node representations. Many pretext tasks, or loss
functions have been designed from distinct pers
Large-scale labeled data are generally required to train deep neural networks in order to obtain better performance in visual feature learning from images or videos for computer vision applications. T
Self-supervised learning aims to learn good representations with unlabeled data.
In this work, instead of designing a new pseudo task for self-supervised learning, we develop a model compression method to compress an already learned, deep self-supervised model (teacher) to a smaller one(student).
Human adaptability relies crucially on learning and merging knowledge from
both supervised and unsupervised tasks: the parents point out few important
concepts, but then the children fill in the gaps
Knowledge distillation of self-supervised vision transformers (vit-sskd) is an important research topic to increase their performance on memory and compute devices.
We show that directly distilling information from the crucial attention mechanism from teacher to student can significantly narrow the performance gap between both.
In this paper, we provide an information-theoretic perspective on variation-invariance-covariance regularization (vicreg) for self-supervised learning and derive a generalization bound for vicreg, providing generalization guarantees for downstream supervised learning tasks.
Next, we relate the vicregobjective to mutual information maximization and use it to highlight the underlying assumptions of the objective.