Detail publikačního výsledku

Large Language Models for Multilingual Previously Fact-Checked Claim Detection

VYKOPAL, I.; PIKULIAK, M.; OSTERMANN, S.; ANIKINA, T.; GREGOR, M.; ŠIMKO, M.

Originální název

Large Language Models for Multilingual Previously Fact-Checked Claim Detection

Anglický název

Large Language Models for Multilingual Previously Fact-Checked Claim Detection

Druh

Stať ve sborníku mimo WoS a Scopus

Originální abstrakt

In our era of widespread false information, human fact-checkers often face the challenge of duplicating efforts when verifying claims that may have already been addressed in other countries or languages. As false information transcends linguistic boundaries, the ability to automatically detect previously fact-checked claims across languages has become an increasingly important task. This paper presents the first comprehensive evaluation of large language models (LLMs) for multilingual previously fact-checked claim detection. We assess seven LLMs across 20 languages in both monolingual and cross-lingual settings. Our results show that while LLMs perform well for high-resource languages, they struggle with low-resource languages. Moreover, translating original texts into English proved to be beneficial for low-resource languages. These findings highlight the potential of LLMs for multilingual previously fact-checked claim detection and provide a foundation for further research on this promising application of LLMs.

Anglický abstrakt

In our era of widespread false information, human fact-checkers often face the challenge of duplicating efforts when verifying claims that may have already been addressed in other countries or languages. As false information transcends linguistic boundaries, the ability to automatically detect previously fact-checked claims across languages has become an increasingly important task. This paper presents the first comprehensive evaluation of large language models (LLMs) for multilingual previously fact-checked claim detection. We assess seven LLMs across 20 languages in both monolingual and cross-lingual settings. Our results show that while LLMs perform well for high-resource languages, they struggle with low-resource languages. Moreover, translating original texts into English proved to be beneficial for low-resource languages. These findings highlight the potential of LLMs for multilingual previously fact-checked claim detection and provide a foundation for further research on this promising application of LLMs.

Autoři

VYKOPAL, I.; PIKULIAK, M.; OSTERMANN, S.; ANIKINA, T.; GREGOR, M.; ŠIMKO, M.

Vydáno

30.10.2025

Nakladatel

Association for Computational Linguistics

Místo

Suzhou, China

ISBN

979-8-8917-6335-7

Strany od

15741

Strany do

15765

Strany počet

25

URL

BibTex

@inproceedings{BUT198601,
  author="Ivan {Vykopal} and  {} and  {} and  {} and Michal {Gregor} and Marián {Šimko}",
  title="Large Language Models for Multilingual Previously Fact-Checked Claim Detection",
  year="2025",
  pages="15741--15765",
  publisher="Association for Computational Linguistics",
  address="Suzhou, China",
  doi="10.18653/v1/2025.findings-emnlp.852",
  isbn="979-8-8917-6335-7",
  url="https://aclanthology.org/2025.findings-emnlp.852/"
}