Title
TabFact: A Large-scale Dataset for Table-based Fact Verification
Abstract
The problem of verifying whether a textual hypothesis holds based on the given evidence, also known as fact verification, plays an important role in the study of natural language understanding and semantic representation. However, existing studies are mainly restricted to dealing with unstructured evidence (e.g., natural language sentences and documents, news, etc), while verification under structured evidence, such as tables, graphs, and databases, remains unexplored. This paper specifically aims to study the fact verification given semi-structured data as evidence. To this end, we construct a large-scale dataset called TabFact with 16k Wikipedia tables as the evidence for 118k human-annotated natural language statements, which are labeled as either ENTAILED or REFUTED. TabFact is challenging since it involves both soft linguistic reasoning and hard symbolic reasoning. To address these reasoning challenges, we design two different models: Table-BERT and Latent Program Algorithm (LPA). Table-BERT leverages the state-of-the-art pre-trained language model to encode the linearized tables and statements into continuous vectors for verification. LPA parses statements into LISP-like programs and executes them against the tables to obtain the returned binary value for verification. Both methods achieve similar accuracy but still lag far behind human performance. We also perform a comprehensive analysis to demonstrate great future opportunities.
Year
Venue
Keywords
2020
ICLR
Fact Verification, Tabular Data, Symbolic Reasoning
DocType
Citations 
PageRank 
Conference
1
0.34
References 
Authors
33
8
Name
Order
Citations
PageRank
Wenhu Chen1459.80
Hongmin Wang211.02
Jianshu Chen388352.94
Yunkai Zhang411.02
H. L. Wang5181.83
Li, Shiyang631.43
Xiyou Zhou751.50
William Yang Wang849359.64