In this paper, we propose an entity-based neural local coherence model which is linguistically more sound than previously proposed neural coherence models. Recent neural coherence models encode the input document using large-scale pretrained language models. Hence their basis for computing local coherence are words and even sub-words. The analysis of their output shows that these models frequently compute coherence on the basis of connections between (sub-)words which, from a linguistic perspective, should not play a role. Still, these models achieve state-of-the-art performance in several end applications. In contrast to these models, we compute coherence on the basis of entities by constraining the input to noun phrases and proper names. This provides us with an explicit representation of the most important items in sentences leading to the notion of focus. This brings our model linguistically in line with pre-neural models of computing coherence. It also gives us better insight into the behaviour of the model thus leading to better explainability. Our approach is also in accord with a recent study (O’Connor and Andreas, 2021), which shows that most usable information is captured by nouns and verbs in transformer-based language models. We evaluate our model on three downstream tasks showing that it is not only linguistically more sound than previous models but also that it outperforms them in end applications.
SEEK ID: https://publications.h-its.org/publications/1471
Research Groups: Natural Language Processing
Publication type: InProceedings
Book Title: In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (ACL)
Publisher: Association for Computational Linguistics
Citation: In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (ACL). Dublin, Ireland, May 2022
Date Published: 22nd May 2022
URL: https://aclanthology.org/2022.acl-long.537.pdf
Registered Mode: manually
Views: 3895
Created: 9th Mar 2022 at 12:36
Last updated: 5th Mar 2024 at 21:24
This item has not yet been tagged.
None