Abstract
The distributed training of machine learning (ML) models presents significant challenges in ensuring data and parameter protection. Privacy-enhancing technologies (PETs) offer a promising initial step towards addressing these concerns, yet achieving confidentiality and differential privacy in distributed learning remains complex. This paper introduces a novel data protection technique tailored for the distributed training of ML models, ensuring compliance with regulatory standards. Our approach utilizes a quantized multi-hash data representation, known as Hash-Comb, combined with randomization to achieve Rényi differential privacy (RDP) for both training data and model parameters. The training protocol is designed to require only the common knowledge of a few hyper-parameters, which are securely shared using multi-party computation protocols. Experimental results demonstrate the effectiveness of our method in preserving both privacy and model accuracy.
| Original language | British English |
|---|---|
| Article number | 107741 |
| Journal | Future Generation Computer Systems |
| Volume | 167 |
| DOIs | |
| State | Published - Jun 2025 |
Keywords
- Confidentiality
- Differential privacy
- Hashing
- Random quantization