TY - JOUR
T1 - Consensus-based distributed support vector machines
AU - Forero, Pedro A.
AU - Cano, Alfonso
AU - Giannakis, Georgios B.
PY - 2010/5/1
Y1 - 2010/5/1
N2 - This paper develops algorithms to train support vector machines when training data are distributed across different nodes, and their communication to a centralized processing unit is prohibited due to, for example, communication complexity, scalability, or privacy reasons. To accomplish this goal, the centralized linear SVM problem is cast as a set of decentralized convex optimization subproblems (one per node) with consensus constraints on the wanted classifier parameters. Using the alternating direction method of multipliers, fully distributed training algorithms are obtained without exchanging training data among nodes. Different from existing incremental approaches, the overhead associated with inter-node communications is fixed and solely dependent on the network topology rather than the size of the training sets available per node. Important generalizations to train nonlinear SVMs in a distributed fashion are also developed along with sequential variants capable of online processing. Simulated tests illustrate the performance of the novel algorithms.1
AB - This paper develops algorithms to train support vector machines when training data are distributed across different nodes, and their communication to a centralized processing unit is prohibited due to, for example, communication complexity, scalability, or privacy reasons. To accomplish this goal, the centralized linear SVM problem is cast as a set of decentralized convex optimization subproblems (one per node) with consensus constraints on the wanted classifier parameters. Using the alternating direction method of multipliers, fully distributed training algorithms are obtained without exchanging training data among nodes. Different from existing incremental approaches, the overhead associated with inter-node communications is fixed and solely dependent on the network topology rather than the size of the training sets available per node. Important generalizations to train nonlinear SVMs in a distributed fashion are also developed along with sequential variants capable of online processing. Simulated tests illustrate the performance of the novel algorithms.1
KW - Distributed data mining
KW - Distributed learning
KW - Distributed optimization
KW - Sensor networks
KW - Support vector machine
UR - http://www.scopus.com/inward/record.url?scp=77953526250&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=77953526250&partnerID=8YFLogxK
M3 - Article
AN - SCOPUS:77953526250
SN - 1532-4435
VL - 11
SP - 1663
EP - 1707
JO - Journal of Machine Learning Research
JF - Journal of Machine Learning Research
ER -