Machine learning models are not static and may need to be retrained on
slightly changed datasets, for instance, with the addition or deletion of a set
of data points. This has many applications, inclu
Machine learning (ML) is becoming a commodity. Numerous ML frameworks and
services are available to data holders who are not ML experts but want to train
predictive models on their data. It is importa
Given the computational cost and technical expertise required to train
machine learning models, users may delegate the task of learning to a service
provider. We show how a malicious learner can plant
Machine learning (ML) models may be deemed confidential due to their
sensitive training data, commercial value, or use in security applications.
Increasingly often, confidential ML models are being de
Adversarial examples causing evasive predictions are widely used to evaluate
and improve the robustness of machine learning models. However, current studies
on adversarial examples focus on supervised
Deleting data from a trained machine learning (ML) model is a critical task
in many applications. For example, we may want to remove the influence of
training points that might be out of date or outli
Making statements about the performance of trained models on tasks involving
new data is one of the primary goals of machine learning, i.e., to understand
the generalization power of a model. Various
This technical manuscript summarises, formalises and extends the link between quantum machine learning and kernel methods by systematically rephrasing quantum models as a kernel method.
It shows that most near-term and fault-tolerant quantum models can be replaced by a general support vector machine whose kernel computes distances between data-encoding quantum states.
Training machine learning models on classical computers is usually a time and compute intensive process. With Moore's law coming to an end and ever increasing demand for large-scale data analysis usin
Correctly quantifying the robustness of machine learning models is a central
aspect in judging their suitability for specific tasks, and thus, ultimately,
for generating trust in the models. We show t
Whose labels should a machine learning (ML) algorithm learn to emulate? For
ML tasks ranging from online comment toxicity to misinformation detection to
medical diagnosis, different groups in society
Good data stewardship requires removal of data at the request of the data's
owner. This raises the question if and how a trained machine-learning model,
which implicitly stores information about its t
The introduction of robust optimisation has pushed the state-of-the-art in
defending against adversarial attacks. However, the behaviour of such
optimisation has not been studied in the light of a fun
The question of answering queries over ML predictions has been gaining
attention in the database community. This question is challenging because the
cost of finding high quality answers corresponds to
We quantitatively investigate how machine learning models leak information
about the individual data records on which they were trained. We focus on the
basic membership inference attack: given a data
We overview methods for an often-overlooked step in the development of machine learning models : building community trust in the algorithms.
Trust is an essential ingredient not just for creating morerobust data analysis techniques, but also for building confidence within the astronomy community to embrace machine learning methods and results.
The vast advances in Machine Learning over the last ten years have been
powered by the availability of suitably prepared data for training purposes.
The future of ML-enabled enterprise hinges on data.
Machine learning models often make basic errors that are easily hidden within vast amounts of data.
We propose a framework that integrates logic-based methods with statistical inference to derive common sense rules from a model s training data without supervision.