Researchers at the Technical University of Darmstadt’s Ubiquitous Knowledge Processing Lab have discovered a technique to enhance the performance of smaller language models (SLMs) in extractive question answering. Led by Rachneet Sachdeva, Martin Tutek, and Iryna Gurevych, the team utilized large language models (LLMs) for data augmentation by generating counterfactual (CF) instances — minimally modified input data — which notably improved out-of-domain (OOD) performance. Their study, “CATfOOD: Counterfactual Augmented Training for Improving Out-of-Domain Performance and Calibration,” indicates that the diversity of CF instances, both in form and content, is key to mitigating spurious correlations and addressing data distribution disparities.
For further actions, you may consider blocking this person and/or reporting abuse
Top comments (0)