Analysis and Implementation of Large-scale Linear RankSVM in Distributed Environments
Date Issued
2016
Date
2016
Author(s)
Huang, Wei-Lun
Abstract
Linear rankSVM is a useful method to quickly produce a baseline model for learning to rank. Although its parallelization has been investigated and implemented on GPU, it may not handle large-scale data sets. In this thesis, we propose a distributed trust region Newton method for training L2-loss linear rankSVM with two kinds of parallelizations. We carefully discuss the techniques for reducing the communication cost and speeding up the computation, and compare both kinds of parallelizations on dense and sparse data sets. Experiments show that our distributed methods are much faster than the single machine method on two kinds of data sets: one with its number of instances much larger than its number of features, and the other is the opposite.
Subjects
Learning to rank
Ranking support vector machines
Large-scale learning
Linear model
Distributed Newton method
Type
thesis
File(s)![Thumbnail Image]()
Loading...
Name
ntu-105-R02922041-1.pdf
Size
23.32 KB
Format
Adobe PDF
Checksum
(MD5):e70328aaa5f8b087f6c71f831974fc1f