Salta alla navigazione principale Salta alla ricerca Salta al contenuto principale

A defense mechanism against label inference attacks in Vertical Federated Learning

  • Marco Arazzi
  • , Serena Nicolazzo
  • , Antonino Nocera

Risultato della ricerca: Contributo su rivistaArticolo in rivistapeer review

Abstract

Vertical Federated Learning (VFL, for short) is a category of Federated Learning that is gaining increasing attention in the context of Artificial Intelligence. According to this paradigm, machine/deep learning models are trained collaboratively among parties with vertically partitioned data. Typically, in a VFL scenario, the labels of the samples are kept private from all parties except the aggregating server, that is, the label owner. However, recent work discovered that by exploiting the gradient information returned by the server to bottom models, with the knowledge of only a small set of auxiliary labels on a very limited subset of training data points, an adversary could infer the private labels. These attacks are known as label inference attacks in VFL. In our work, we propose a novel framework called KDk (knowledge distillation with k-anonymity) that combines knowledge distillation and k-anonymity to provide a defense mechanism against potential label inference attacks in a VFL scenario. Through an exhaustive experimental campaign, we demonstrate that by applying our approach, the performance of the analyzed label inference attacks decreases consistently, even by more than 60%, maintaining the accuracy of the whole VFL almost unaltered.

Lingua originaleInglese
Numero di articolo129476
RivistaNeurocomputing
Volume624
DOI
Stato di pubblicazionePubblicato - 1 apr 2025
Pubblicato esternamente

Fingerprint

Entra nei temi di ricerca di 'A defense mechanism against label inference attacks in Vertical Federated Learning'. Insieme formano una fingerprint unica.

Cita questo