Skip to content

janekh24/anhalten

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ANHALTEN: Cross-Lingual Transfer for German Token-Level Reference-Free Hallucination Detection

Authors: Janek Herrlein, Chia-Chien Hung, Goran Glavaš

ACL 2024. SRW: https://aclanthology.org/2024.acl-srw.18

Introduction

Research on token-level reference-free hallucination detection has predominantly focused on English, primarily due to the scarcity of robust datasets in other languages. This has hindered systematic investigations into the effectiveness of cross-lingual transfer for this important NLP application. To address this gap, we introduce ANHALTEN, a new evaluation dataset that extends the English hallucination detection dataset to German. To the best of our knowledge, this is the first work that explores cross-lingual transfer for token-level reference-free hallucination detection. ANHALTEN contains gold annotations in German that are parallel (i.e., directly comparable to the original English instances). We benchmark several prominent cross-lingual transfer approaches, demonstrating that larger context length leads to better hallucination detection in German, even without succeeding context. Importantly, we show that the sample-efficient few-shot transfer is the most effective approach in most setups. This highlights the practical benefits of minimal annotation effort in the target language for reference-free hallucination detection.

Citation

If you use any source codes, or datasets included in this repo in your work, please cite the following paper:

@inproceedings{herrlein-etal-2024-anhalten,
    title = "{ANHALTEN}: Cross-Lingual Transfer for {G}erman Token-Level Reference-Free Hallucination Detection",
    author = "Herrlein, Janek  and
      Hung, Chia-Chien  and
      Glava{\v{s}}, Goran",
    booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop)",
    month = aug,
    year = "2024",
    address = "Bangkok, Thailand",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2024.acl-srw.18",
    pages = "186--194"
}

Pretrained Models

The pre-trained models can be easily loaded using huggingface Transformers or Adapter-Hub adapter-transformers library. Following pre-trained versions are supported:

  • bert-base-multilingual-cased: mBERT
  • xlm-roberta-base: XLM-R
  • en/wiki@ukp: Adapter trained on English Wiki
  • de/wiki@ukp: Adapter trained on German Wiki

The scripts for downstream tasks are mainly modified from here, where there might be slight version differences of the packages, which are noted down in the requirements.txt file.

Structure

This repository is currently under the following structure:

.
└── Code
└── Data
└── README.md

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published