Technical perspective of TURL: Table understanding through representation learning

Papotti, Paolo
ACM SIGMOD Record, Volume 51, Issue 1, March 2022

Several efforts aim at representing tabular data with neural models for supporting target applications at the intersection of natural language processing (NLP) and databases (DB) [1-3]. The goal is to extend to structured data the recent neural architectures, which achieve state of the art results in NLP applications. Language models (LMs) are usually pre-trained with unsupervised tasks on a large text corpus. The output LM is then fine-tuned on a variety of downstream tasks with a small set of specific examples. This process has many advantages, because the LM contains information about textual structure and content, which are used by the target application without manually defining features.


DOI
Type:
Journal
Date:
2022-06-01
Department:
Data Science
Eurecom Ref:
6906
Copyright:
© ACM, 2022. This is the author's version of the work. It is posted here by permission of ACM for your personal use. Not for redistribution. The definitive version was published in ACM SIGMOD Record, Volume 51, Issue 1, March 2022 https://doi.org/10.1145/3542700.3542708
See also:

PERMALINK : https://www.eurecom.fr/publication/6906