信息网络安全 ›› 2026, Vol. 26 ›› Issue (2): 236-250.doi: 10.3969/j.issn.1671-1122.2026.02.005

• 学术研究 • 上一篇    下一篇

面向分布式学习的多密钥同态加密与差分隐私融合方案

王腾(), 樊坤渭, 张瑶   

  1. 西安邮电大学网络空间安全学院西安 710121
  • 收稿日期:2025-04-15 出版日期:2026-02-10 发布日期:2026-02-23
  • 通讯作者: 王腾 wangteng@xupt.edu.cn
  • 作者简介:王腾(1995—),女,陕西,副教授,博士,主要研究方向为数据安全与隐私保护|樊坤渭(2000—),男,陕西,硕士研究生,主要研究方向为机器学习、隐私保护|张瑶(2001—),女,陕西,硕士研究生,主要研究方向为实时数据流隐私保护
  • 基金资助:
    国家自然科学基金(62102311);陕西省科学技术协会青年人才托举计划(20240116);陕西省重点研发计划(2025CY-YBXM-069)

A Fusion Scheme of Multi-Key Homomorphic Encryption and Differential Privacy for Distributed Learning

WANG Teng(), FAN Kunwei, ZHANG Yao   

  1. School of Cyberspace Security, Xi’an University of Posts and Telecommunications, Xi’an 710121, China
  • Received:2025-04-15 Online:2026-02-10 Published:2026-02-23

摘要:

在大数据时代,机器学习领域的数据隐私保护愈发重要。在多方学习场景中,攻击者能够根据梯度和模型参数等信息,反向推导出原始数据特征。此外,部分参与方可能为私利而相互串通,共享本应保密的数据,从而破坏多方学习的公平性和隐私保护需求。为解决上述问题,文章提出面向分布式学习的多密钥同态加密与差分隐私融合方案,即PrivMPL方案,其核心目标是在确保数据隐私安全的前提下,实现高效的模型训练。在该方案中,本地客户端使用聚合公钥加密更新后的模型参数,解密过程需要所有数据使用者协同完成。服务器通过对聚合参数添加高斯噪声实现差分隐私。该方案有效防止多方训练过程中因共享信息导致的隐私泄露,并且对数据使用者与服务器之间的共谋具有鲁棒性。为验证PrivMPL方案的有效性,将PrivMPL方案与基于Paillier的同态加密多方学习方案进行对比,以模型准确率作为评估指标。实验结果表明,PrivMPL方案在模型准确率方面有显著提升,进一步证明了该方案在数据隐私保护和模型性能等方面的优势。

关键词: 机器学习即服务, 多方学习, 差分隐私, 多密钥同态加密

Abstract:

In the era of big data, data privacy protection in the field of machine learning has become increasingly important. In multi-party learning scenarios, attackers can reverse-engineer original data features from information such as gradients and model parameters. Moreover, some participants may collude for personal gain, sharing data that should remain confidential, thereby undermining the fairness and privacy requirements of multi-party learning. To address these issues, this paper proposed a fusion scheme of multi-key homomorphic encryption and differential privacy for distributed learning, namely the PrivMPL scheme, whose core objective was to achieve efficient model training while ensuring data privacy security. In this scheme, local clients used an aggregated public key to encrypt updated model parameters, and the decryption process required collaborative participation from all data users. The server achieved differential privacy by adding Gaussian noise to the aggregated parameters. The scheme effectively prevented privacy leakage caused by information sharing during multi-party training and was robust against collusion between data users and the server. To validate the effectiveness of the PrivMPL scheme, it is compared with a Paillier-based homomorphic encryption multi-party learning approach, using model accuracy as the evaluation metric. Experimental results show that the PrivMPL scheme achieves a significant improvement in model accuracy, further demonstrating its advantages in data privacy protection and model performance.

Key words: machine learning as a service, multi-party learning, differential privacy, multi-key homomorphic encryption

中图分类号: