Title
A sparse {\varvec{L}}_{2}-regularized support vector machines for efficient natural language learning
Abstract
Linear kernel support vector machines (SVMs) using either \(L_{1}\)-norm or \(L_{2}\)-norm have emerged as an important and wildly used classification algorithm for many applications such as text chunking, part-of-speech tagging, information retrieval, and dependency parsing. \(L_{2}\)-norm SVMs usually provide slightly better accuracy than \(L_{1}\)-SVMs in most tasks. However, \(L_{2}\)-norm SVMs produce too many near-but-nonzero feature weights that are highly time-consuming when computing nonsignificant weights. In this paper, we present a cutting-weight algorithm to guide the optimization process of the \(L_{2}\)-SVMs toward a sparse solution. Before checking the optimality, our method automatically discards a set of near-but-nonzero feature weight. The final objects can then be achieved when the objective function is met by the remaining features and hypothesis. One characteristic of our cutting-weight algorithm is that it requires no changes in the original learning objects. To verify this concept, we conduct the experiments using three well-known benchmarks, i.e., CoNLL-2000 text chunking, SIGHAN-3 Chinese word segmentation, and Chinese word dependency parsing. Our method achieves 1–10 times feature parameter reduction rates in comparison with the original \(L_{2}\)-SVMs, slightly better accuracy with a lower training time cost. In terms of run-time efficiency, our method is reasonably faster than the original \(L_{2}\)-regularized SVMs. For example, our sparse \(L_{2}\)-SVMs is 2.55 times faster than the original \(L_{2}\)-SVMs with the same accuracy.
Year
DOI
Venue
2014
10.1007/s10115-013-0615-0
Knowledge and Information Systems
Keywords
Field
DocType
L2-regularization, Support vector machines, Machine learning, Text chunking, Dependency parsing, Chinese word segmentation
Kernel (linear algebra),Data mining,Computer science,Support vector machine,Dependency grammar,Text segmentation,Natural language,Regularization (mathematics),Artificial intelligence,Chunking (psychology),Machine learning,Parameter reduction
Journal
Volume
Issue
ISSN
39
2
0219-3116
Citations 
PageRank 
References 
1
0.36
22
Authors
1
Name
Order
Citations
PageRank
Yu-Chieh Wu124723.16