Loading…
A secure distributed machine learning protocol against static semi-honest adversaries
Machine learning has been successfully applied to various fields over the last few years. However, it still faces two critical challenges. On one hand, the concern for the security issue in machine learning is increased. On the other hand, data exists in the form of isolated islands across different...
Saved in:
Published in: | Applied soft computing 2021-04, Vol.102, p.107095, Article 107095 |
---|---|
Main Authors: | , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Machine learning has been successfully applied to various fields over the last few years. However, it still faces two critical challenges. On one hand, the concern for the security issue in machine learning is increased. On the other hand, data exists in the form of isolated islands across different organizations. In this work, we focus on the privacy-preserving issue on a non-parametric machine learning algorithmize, i.e., the k Nearest Neighbor Classification (k NNC) in which, training data are split vertically among multiple servers. We propose a novel protocol that is secure against static semi-honest adversaries. In specific, the clients can obtain the label of his/her query without disclosing the servers’ data, the client’s query, and the client’s output to others. We use the state-of-the-art lattice-based fully homomorphic encryption to realize the privacy-preserving distance computation. In order to protect data access patterns, permutation technique and Oblivious Transfer are used in the top-k selection phase. We proved the security via the simulation paradigm. Meanwhile, we implemented our protocol and performed extensive experiments. Results show that our protocol performs well, especially in a large-width environment. Compared to the existing solution, our protocol leaks no information about the participants’ private input and output in both centralized and distributed architectures. Meanwhile, our protocol runs faster than existing solutions.
•A novel secure distributed k Nearest Neighbor classification protocol is proposed.•The protocol works in a distributed architecture, where the training data are split vertically among multiple servers.•The protocol accepts all types of decision rules.•The protocol is secure against static semi-honest adversaries.•The implementation performs faster than some existing protocols. |
---|---|
ISSN: | 1568-4946 1872-9681 |
DOI: | 10.1016/j.asoc.2021.107095 |