















Guided by uncertainty
Machine learning
A new two-stage method addresses challenges in the natural language processing of long texts using transformers with self-attention mechanisms.
Uncertainty Guided Global Memory Improves Multi-Hop Question Answering
Submitted
Uncertainty Guided Global Memory Improves Multi-Hop Question Answering Transformers have become the gold standard for many natural language processing tasks, however, models with self-attention mechanisms struggle to process long sequences due to their quadratic complexity. Therefore, processing long texts remains a challenge. To address this issue, we propose a two-stage method that first collects relevant information over the entire document and then combines it with local context to solve the task. Our experimental results show that fine-tuning a pre-trained model with memory-augmented input, including the least uncertain global elements, improves the model's performance on multi-hop question answering task compared to the baseline. We also found that the content of the global memory correlates with the supporting facts required for the correct answer.
Submitted