Authors: Janek Herrlein, Chia-Chien Hung, Goran Glavaš
ACL 2024. SRW: https://aclanthology.org/2024.acl-srw.18
Research on token-level reference-free hallucination detection has predominantly focused on English, primarily due to the scarcity of robust datasets in other languages. This has hindered systematic investigations into the effectiveness of cross-lingual transfer for this important NLP application. To address this gap, we introduce ANHALTEN, a new evaluation dataset that extends the English hallucination detection dataset to German. To the best of our knowledge, this is the first work that explores cross-lingual transfer for token-level reference-free hallucination detection. ANHALTEN contains gold annotations in German that are parallel (i.e., directly comparable to the original English instances). We benchmark several prominent cross-lingual transfer approaches, demonstrating that larger context length leads to better hallucination detection in German, even without succeeding context. Importantly, we show that the sample-efficient few-shot transfer is the most effective approach in most setups. This highlights the practical benefits of minimal annotation effort in the target language for reference-free hallucination detection.
If you use any source codes, or datasets included in this repo in your work, please cite the following paper:
@inproceedings{herrlein-etal-2024-anhalten, title = "{ANHALTEN}: Cross-Lingual Transfer for {G}erman Token-Level Reference-Free Hallucination Detection", author = "Herrlein, Janek and Hung, Chia-Chien and Glava{\v{s}}, Goran", booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop)", month = aug, year = "2024", address = "Bangkok, Thailand", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2024.acl-srw.18", pages = "186--194" }
The pre-trained models can be easily loaded using huggingface Transformers or Adapter-Hub adapter-transformers library. Following pre-trained versions are supported:
bert-base-multilingual-cased
: mBERTxlm-roberta-base
: XLM-Ren/wiki@ukp
: Adapter trained on English Wikide/wiki@ukp
: Adapter trained on German Wiki
The scripts for downstream tasks are mainly modified from here, where there might be slight version differences of the packages, which are noted down in the requirements.txt
file.
This repository is currently under the following structure:
.
└── Code
└── Data
└── README.md