ICNLSP 2024: Semantically Enriched Text Generation for QA through Dense Paraphrasing
ฝัง
- เผยแพร่เมื่อ 23 ธ.ค. 2024
- Semantically Enriched Text Generation for QA through Dense Paraphrasing
By: Timothy Obiso, Bingyang Ye, Kyeongmin Rim, James Pustejovsky
Brandeis University
7th International Conference on Natural Language and Speech Processing.
icnlsp.org/202...
Abstract:
Large language models (LLMs) are very effective at extractive language tasks such as QA. While LLMs can improve their performance on these tasks through increases in model size (via massive pretraining) and/or iterative on-the-job training (one-shot, few-shot, chain-of-thought), we explore what other less resource-intensive and more efficient types of data augmentation can be applied to obtain similar boosts in performance. We define multiple forms of Dense Paraphrasing (DP) and obtain DP-enriched versions of different contexts. We demonstrate that performing QA using these semantically enriched contexts leads to increased performance on models of various sizes and across task domains, without needing to increase model size.