Current approaches for detecting text reuse do not focus on recontextualization, i.e., how the new context(s) of a reused text differs from its original context(s). In this paper, we propose a novel framework called TRoTR that relies on the notion of topic relatedness for evaluating the diachronic change of context in which text is reused. TRoTR includes two NLP tasks: TRiC and TRaC. TRiC is designed to evaluate the topic relatedness between a pair of recontextualizations. TRaC is designed to evaluate the overall topic variation within a set of recontextualizations. We also provide a curated TRoTR benchmark of biblical text reuse, human-annotated with topic relatedness. The benchmark exhibits an inter-annotator agreement of.811. We evaluate multiple, established SBERT models on the TRoTR tasks and find that they exhibit greater sensitivity to textual similarity than topic relatedness. Our experiments show that fine-tuning these models can mitigate such a kind of sensitivity.
TRoTR: A Framework for Evaluating the Recontextualization of Text / F. Periti, P. Cassotti, S. Montanelli, N. Tahmasebi, D. Schlechtweg - In: 2024 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference[s.l] : Association for Computational Linguistics, 2024. - ISBN 979-8-89176-164-3. - pp. 13972-13990 (( convegno Conference on Empirical Methods in Natural Language Processing tenutosi a Miami nel 2024 [10.18653/v1/2024.emnlp-main.774].
TRoTR: A Framework for Evaluating the Recontextualization of Text
F. Periti;S. Montanelli;
2024
Abstract
Current approaches for detecting text reuse do not focus on recontextualization, i.e., how the new context(s) of a reused text differs from its original context(s). In this paper, we propose a novel framework called TRoTR that relies on the notion of topic relatedness for evaluating the diachronic change of context in which text is reused. TRoTR includes two NLP tasks: TRiC and TRaC. TRiC is designed to evaluate the topic relatedness between a pair of recontextualizations. TRaC is designed to evaluate the overall topic variation within a set of recontextualizations. We also provide a curated TRoTR benchmark of biblical text reuse, human-annotated with topic relatedness. The benchmark exhibits an inter-annotator agreement of.811. We evaluate multiple, established SBERT models on the TRoTR tasks and find that they exhibit greater sensitivity to textual similarity than topic relatedness. Our experiments show that fine-tuning these models can mitigate such a kind of sensitivity.| File | Dimensione | Formato | |
|---|---|---|---|
|
2024.emnlp-main.774.pdf
accesso aperto
Tipologia:
Publisher's version/PDF
Dimensione
379.93 kB
Formato
Adobe PDF
|
379.93 kB | Adobe PDF | Visualizza/Apri |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.




