MaskCRYPT: Federated Learning With Selective Homomorphic Encryption

Published:

MaskCRYPT: Federated Learning With Selective Homomorphic Encryption

This post is based on the paper MaskCRYPT: Federated Learning With Selective Homomorphic Encryption.
이 글은 MaskCRYPT: Federated Learning With Selective Homomorphic Encryption 논문을 기반으로 정리한 내용이다.

Overview

Summary:
A review of MaskCRYPT, which proposes selective homomorphic encryption to balance privacy and efficiency in federated learning.

🔑 Research Question:

  • Do we need to encrypt all model weights to ensure privacy?

⚙️ Key Mechanism: Selective Homomorphic Encryption

  • Uses gradient-based importance to identify which parameters should be encrypted
  • Each client generates a local mask based on importance scores
  • The server aggregates these into a global Mask Consensus
  • Only selected parameters are encrypted, while others are aggregated in plaintext

📊 Main Results:

  • Encrypting as little as 1% of parameters can defend against membership inference and reconstruction attacks
  • Reduces communication overhead by up to 4.15×
  • Improves wall-clock training time
  • Maintains accuracy comparable to non-encrypted training

⚠️ Limitations / Open Questions:

  • Requires communication of local importance scores
  • Fairness and correctness of Mask Consensus must be ensured
  • Evaluated only on moderate-scale models and datasets

💡 Insight:
MaskCRYPT suggests that full encryption may be unnecessary in practical federated learning systems, and that selectively protecting sensitive components can provide a better trade-off between security and efficiency.


Slides:
PDF (Korean) Download