Title
Kernelized Hashcode Representations for Relation Extraction
Abstract
Kernel methods have produced state-of-the-art results for a number of NLP tasks such as relation extraction, but stiffer from poor scalability due to the high cost of computing kernel similarities between natural language structures. A recently proposed technique, kernelized locality-sensitive hashing (KLSH), can significantly reduce the computational cost, but is only applicable to classifiers operating on kNN graphs. Here we propose to use random subspaces of KLSH codes for efficiently constructing an explicit representation of NLP structures suitable for general classification methods. Further, we propose an approach for optimizing the KLSH model for classification problems by maximizing an approximation of mutual information between the KLSH codes (feature vectors) and the class labels. We evaluate the proposed approach on biomedical relation extraction datasets, and observe significant and robust improvements in accuracy w.r.t. state-of-the-art classifiers, along with drastic (orders-of-magnitude) speedup compared to conventional kernel methods.
Year
DOI
Venue
2019
10.1609/aaai.v33i01.33016431
AAAI
Field
DocType
Volume
Kernel (linear algebra),Feature vector,Computer science,Mutual information,Artificial intelligence,Java hashCode,Kernel method,Machine learning,Relationship extraction,Scalability,Speedup
Conference
33
Citations 
PageRank 
References 
0
0.34
0
Authors
6
Name
Order
Citations
PageRank
Sahil Garg193.88
Aram Galstyan2103394.05
Greg Ver Steeg324332.99
irina rish491281.78
Guillermo A. Cecchi519934.56
Shuyang Gao6275.23