Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Deep learning models, while achieving remarkable performances, are vulnerable
to membership inference attacks (MIAs). Although various defenses have been
proposed, there is still substantial room for improvement in the
privacy-utility trade-off. In this work, we introduce a novel defense framework
against MIAs by leveraging generative models. The key intuition of our defense
is to remove the differences between member and non-member inputs, which is
exploited by MIAs, by re-generating input samples before feeding them to the
target model. Therefore, our defense, called DIFFENCE, works pre inference,
which is unlike prior defenses that are either training-time or post-inference
time.
A unique feature of DIFFENCE is that it works on input samples only, without
modifying the training or inference phase of the target model. Therefore, it
can be cascaded with other defense mechanisms as we demonstrate through
experiments. DIFFENCE is designed to preserve the model's prediction labels for
each sample, thereby not affecting accuracy. Furthermore, we have empirically
demonstrated it does not reduce the usefulness of confidence vectors. Through
extensive experimentation, we show that DIFFENCE can serve as a robust
plug-n-play defense mechanism, enhancing membership privacy without
compromising model utility. For instance, DIFFENCE reduces MIA accuracy against
an undefended model by 15.8\% and attack AUC by 14.0\% on average across three
datasets, all without impacting model utility. By integrating DIFFENCE with
prior defenses, we can achieve new state-of-the-art performances in the
privacy-utility trade-off. For example, when combined with the state-of-the-art
SELENA defense it reduces attack accuracy by 9.3\%, and attack AUC by 10.0\%.
DIFFENCE achieves this by imposing a negligible computation overhead, adding
only 57ms to the inference time per sample processed on average.