2019
DOI: 10.48550/arxiv.1907.07735
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning Privately over Distributed Features: An ADMM Sharing Approach

Yaochen Hu,
Peng Liu,
Linglong Kong
et al.

Abstract: Distributed machine learning has been widely studied in order to handle exploding amount of data. In this paper, we study an important yet less visited distributed learning problem where features are inherently distributed or vertically partitioned among multiple parties, and sharing of raw data or model parameters among parties is prohibited due to privacy concerns. We propose an ADMM sharing framework to approach risk minimization over distributed features, where each party only needs to share a single value… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(14 citation statements)
references
References 20 publications
0
14
0
Order By: Relevance
“…As a result, the distributed algorithms developed for (P) can be applied to solving both problems (4) and (7).…”
Section: Application Examplesmentioning
confidence: 99%
See 2 more Smart Citations
“…As a result, the distributed algorithms developed for (P) can be applied to solving both problems (4) and (7).…”
Section: Application Examplesmentioning
confidence: 99%
“…The cross-entropy loss function [38] is applied in the last layer. Then, the proposed PDC algorithm and IPDC algorithm are applied to train the classification NN, respectively, by solving problem (7). For the PDC/IPDC algorithms, it is set that p = 0.01, β = 0.01, ρ ∈ {100, 10, 1, 10 −1 } and α ∈ {10 −4 , 10 −3 , 10 −2 , 10 −1 }.…”
Section: Distributed Neural Networkmentioning
confidence: 99%
See 1 more Smart Citation
“…Most studies (e.g., [24,25,39,47,57,70]) in VFL focus on training and simply assumes record linkage has been done (i.e., implicit exact linkage on record ID), which is impractical since most real-world federated datasets are unlinked. Some approaches exactly link the identifiers by exact PPRL [9] or private set intersection (PSI) [9,36,37,49,62].…”
Section: Related Workmentioning
confidence: 99%
“…However, they also focus only on the most similar identifiers and assume there is a one-to-one mapping between the data records of two parties, which is not always true in practice. Current VFL frameworks support various machine learning models including linear regression [16], logistic regression [24], support vector machine [35], gradient boosting decision trees [9,62]. FDML [25] supports neural networks but it requires all the parties to hold labels.…”
Section: Related Workmentioning
confidence: 99%