Loading…
Secure SVM Training Over Vertically-Partitioned Datasets Using Consortium Blockchain for Vehicular Social Networks
Machine learning (ML) techniques are expected to be used for specific applications in Vehicular Social Networks (VSNs). Support vector machine (SVM) is one of the typical ML methods and widely used for its high efficiency. Due to the limitation of data sources, the data collected by different entiti...
Saved in:
Published in: | IEEE transactions on vehicular technology 2020-06, Vol.69 (6), p.5773-5783 |
---|---|
Main Authors: | , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Machine learning (ML) techniques are expected to be used for specific applications in Vehicular Social Networks (VSNs). Support vector machine (SVM) is one of the typical ML methods and widely used for its high efficiency. Due to the limitation of data sources, the data collected by different entities usually contain attributes that are quite different. However, in some real-world scenarios, when training an SVM classifier, many entities face the same problem that they are lacking in data with adequate attributes. Thus multiple entities are required to share data to combine a dataset with diverse attributes and then jointly train a comprehensive classifier. However, data privacy concerns are raised because of data sharing. To sovle the problem, we propose a privacy-preserving SVM classifier training scheme over vertically-partitioned datasets posessed by multiple data providers. In our scheme, we utilize consortium blockchain and threshold homomorphic cryptosystem to establish a secure SVM classifier training platform without a trusted third-party. We keep lots of training operations locally over original data and necessary interactions between participants are protected by the threshold Paillier and consortium blockchain. Security analysis proves that our scheme can preserve the privacy of the original data and the training intermediate values. Extensive experiments indicate that our scheme has high efficiency and no accuracy loss. |
---|---|
ISSN: | 0018-9545 1939-9359 |
DOI: | 10.1109/TVT.2019.2957425 |