Title
Beyond Word Importance: Contextual Decomposition to Extract Interactions from LSTMs.
Abstract
The driving force behind the recent success of LSTMs has been their ability to learn complex and non-linear relationships. Consequently, our inability to describe these relationships has led to LSTMs being characterized as black boxes. To this end, we introduce contextual decomposition (CD), an interpretation algorithm for analysing individual predictions made by standard LSTMs, without any changes to the underlying model. By decomposing the output of a LSTM, CD captures the contributions of combinations of words or variables to the final prediction of an LSTM. On the task of sentiment analysis with the Yelp and SST data sets, we show that CD is able to reliably identify words and phrases of contrasting sentiment, and how they are combined to yield the LSTMu0027s final prediction. Using the phrase-level labels in SST, we also demonstrate that CD is able to successfully extract positive and negative negations from an LSTM, something which has not previously been done.
Year
Venue
Field
2018
ICLR
Data set,Negation,Computer science,Sentiment analysis,Artificial intelligence,Natural language processing,Black box,Machine learning
DocType
Volume
Citations 
Journal
abs/1801.05453
9
PageRank 
References 
Authors
0.48
12
3
Name
Order
Citations
PageRank
W. James Murdoch1322.61
Peter J. Liu226912.28
Bin Yu31984241.03