Will data Shapley values be used to efficiently detect poisoned samples?
9
260Ṁ369
Dec 31
73%
chance

Data Shapley values https://arxiv.org/abs/1904.02868 have been introduced as a way to measure how valuable each training sample is to a machine learning model. This allows to sift through training samples, identifying the most/least valuable ones according to a suitable metric.

In the context of data poisoning, where some samples in the training data have been intentionally manipulated by an adversary to reduce the performance of a model trained on them, will data Shapley values prove pivotal to an efficient strategy to counter poisoning?

My judgment of the publicly available literature plus discussion with colleagues. Resolves at the end of 2025.

Get
Ṁ1,000
to start trading!
Sort by:

iiuc, data poisoning attacks don't need to make a model learn any less, just learn an incorrect output for certain inputs. If this is done sneakily enough, it won't even show up on standard model evaluation metrics. A well-poisoned sample might increase model performance overall (by some objective metric), so it seems hard to counter it like this.

@retr0id Maybe the question is too generic… mmh

bought Ṁ10 NO

Hmm, calculating the shapley value of data seems like an interesting idea. However, won't a poisoned sample just read identically to a sparse sample? My educated guess is that sparse samples are actually quite valuable to reduce the error.

@VitorBosshard I am not sure I understand what you mean by sparse sample in this context?

@mariopasquato A sample that says something true about the world, but isn't repeated thousands of times in the data.

@VitorBosshard I see. Assuming that poisoned samples can be detected upon closer inspection but it’s too costly to inspect all training samples, this would still be valuable

© Manifold Markets, Inc.TermsPrivacy