Title
ConsistSum: Unsupervised Opinion Summarization with the Consistency of Aspect, Sentiment and Semantic
Abstract
ABSTRACTUnsupervised opinion summarization techniques are designed to condense the review data and summarize informative and salient opinions in the absence of golden references. Existing dominant methods generally follow a two-stage framework: first creating the synthetic "review-summary" paired datasets and then feeding them into the generative summary model for supervised training. However, these methods mainly focus on semantic similarity in synthetic dataset creation, ignoring the consistency of aspects and sentiments in synthetic pairs. Such inconsistency also brings a gap to the training and inference of the summarization model. To alleviate this problem, we propose ConsistSum, an unsupervised opinion summarization method devoting to capture the consistency of aspects and sentiment between reviews and summaries. Specifically, ConsistSum first extracts the preliminary "review-summary" pairs from the raw corpus by evaluating the distance of aspect distribution and sentiment distribution. Then, we refine the preliminary summary with the constrained Metropolis-Hastings sampling to produce highly consistent synthetic datasets. In the summarization phase, we adopt the generative model T5 as the summarization model. T5 is fine-tuned for the opinion summarization task by incorporating the loss of predicting aspect and opinion distribution. Experimental results on two benchmark datasets, $i.e.$, Yelp and Amazon, demonstrate the superior performance of ConsistSum over the state-of-the-art baselines.
Year
DOI
Venue
2022
10.1145/3488560.3498463
WSDM
Keywords
DocType
Citations 
opinion summarization, unsupervised method, consistency enhancement
Conference
0
PageRank 
References 
Authors
0.34
8
4
Name
Order
Citations
PageRank
Wenjun Ke101.01
Jinhua Gao2466.94
Huawei Shen373961.40
Xueqi Cheng43148247.04