A note on the decomposition methods for support vector regression
Resource
Neural Computation 14 (6): 1267-1281
Journal
Neural Computation
Journal Volume
14
Journal Issue
6
Pages
1267-1281
Date Issued
2002
Author(s)
Liao, Shuo-Peng
Lin, Hsuan-Tien
Lin, Chih-JenLiao, S.-P.
Lin, H.-T.
Lin, C.-J.
Abstract
The dual formulation of support vector regression involves two closely related sets of variables. When the decomposition method is used, many existing approaches use pairs of indices from these two sets as the working set. Basically, they select a base set first and then expand it so all indices are pairs. This makes the implementation different from that for support vector classification. In addition, a larger optimization subproblem has to be solved in each iteration. We provide theoretical proofs and conduct experiments to show that using the base set as the working set leads to similar convergence (number of iterations). Therefore, by using a smaller working set while keeping a similar number of iterations, the program can be simpler and more efficient.
Other Subjects
article
Type
journal article
File(s)![Thumbnail Image]()
Loading...
Name
13.pdf
Size
213.66 KB
Format
Adobe PDF
Checksum
(MD5):6cc09f6b30075335afd43e6aab02acf8