Towards Robust Token Embeddings for Extractive Question Answering

  • Xun Yao
  • , Junlong Ma
  • , Xinrong Hu
  • , Jie Yang
  • , Yi Guo
  • , Junping Liu

Research output: Chapter in Book / Conference PaperChapterpeer-review

3 Citations (Scopus)

Abstract

Extractive Question Answering (EQA) tasks have gained intensive attention in recent years, while Pre-trained Language Models (PLMs) have been widely adopted for encoding purposes. Yet, PLMs typically take as initial input token embeddings and rely on attention mechanisms to extract contextual representations. In this paper, a simple yet comprehensive framework, termed perturbation for alignment (PFA), is proposed to investigate variations towards token embeddings. A robust encoder is further formed being tolerant against the embedding variation and hence beneficial to subsequent EQA tasks. Specifically, PFA consists of two general modules, including the embedding perturbation (a transformation to produce embedding variations) and the semantic alignment (to ensure the representation similarity from original and perturbed embeddings). Furthermore, the framework is flexible to allow several alignment strategies with different interpretations. Our framework is evaluated on four highly-competitive EQA benchmarks, and PFA consistently improves state-of-the-art models.
Original languageEnglish
Title of host publicationWeb Information Systems Engineering – WISE 2023: 24th International Conference, Proceedings
EditorsFeng Zhang, Hua Wang, Mahmoud Barhamgi, Lu Chen, Rui Zhou
Place of PublicationSingapore
PublisherSpringer Nature Singapore
Pages82-96
Number of pages15
ISBN (Print)9789819972531
DOIs
Publication statusPublished - 2023

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume14306 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Keywords

  • Contextual representation
  • Divergence
  • Extractive Question Answering
  • Token embedding
  • Wasserstein distances

Fingerprint

Dive into the research topics of 'Towards Robust Token Embeddings for Extractive Question Answering'. Together they form a unique fingerprint.

Cite this