TY - CHAP
T1 - Towards Robust Token Embeddings for Extractive Question Answering
AU - Yao, Xun
AU - Ma, Junlong
AU - Hu, Xinrong
AU - Yang, Jie
AU - Guo, Yi
AU - Liu, Junping
PY - 2023
Y1 - 2023
N2 - Extractive Question Answering (EQA) tasks have gained intensive attention in recent years, while Pre-trained Language Models (PLMs) have been widely adopted for encoding purposes. Yet, PLMs typically take as initial input token embeddings and rely on attention mechanisms to extract contextual representations. In this paper, a simple yet comprehensive framework, termed perturbation for alignment (PFA), is proposed to investigate variations towards token embeddings. A robust encoder is further formed being tolerant against the embedding variation and hence beneficial to subsequent EQA tasks. Specifically, PFA consists of two general modules, including the embedding perturbation (a transformation to produce embedding variations) and the semantic alignment (to ensure the representation similarity from original and perturbed embeddings). Furthermore, the framework is flexible to allow several alignment strategies with different interpretations. Our framework is evaluated on four highly-competitive EQA benchmarks, and PFA consistently improves state-of-the-art models.
AB - Extractive Question Answering (EQA) tasks have gained intensive attention in recent years, while Pre-trained Language Models (PLMs) have been widely adopted for encoding purposes. Yet, PLMs typically take as initial input token embeddings and rely on attention mechanisms to extract contextual representations. In this paper, a simple yet comprehensive framework, termed perturbation for alignment (PFA), is proposed to investigate variations towards token embeddings. A robust encoder is further formed being tolerant against the embedding variation and hence beneficial to subsequent EQA tasks. Specifically, PFA consists of two general modules, including the embedding perturbation (a transformation to produce embedding variations) and the semantic alignment (to ensure the representation similarity from original and perturbed embeddings). Furthermore, the framework is flexible to allow several alignment strategies with different interpretations. Our framework is evaluated on four highly-competitive EQA benchmarks, and PFA consistently improves state-of-the-art models.
KW - Contextual representation
KW - Divergence
KW - Extractive Question Answering
KW - Token embedding
KW - Wasserstein distances
UR - https://www.scopus.com/pages/publications/85175977894
U2 - 10.1007/978-981-99-7254-8_7
DO - 10.1007/978-981-99-7254-8_7
M3 - Chapter
AN - SCOPUS:85175977894
SN - 9789819972531
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 82
EP - 96
BT - Web Information Systems Engineering – WISE 2023: 24th International Conference, Proceedings
A2 - Zhang, Feng
A2 - Wang, Hua
A2 - Barhamgi, Mahmoud
A2 - Chen, Lu
A2 - Zhou, Rui
PB - Springer Nature Singapore
CY - Singapore
ER -