This repository contains code to the paper Evaluating Webcam-based Gaze Data as an Alternative for Human Rationale Annotations accepted to LREC-COLING 2024.
Please refer to the paper for further details.
In order to fine-tune the models on XQuAD (excl. the sentences from the eye-tracking corpus).
You can run finetuning.py
followed by extract_lrp_relevance.py
and extract_attention.py
. We recommend running this on a GPU.
For instance like this, where id
refers to a chosen integer to separate multiple runs with the same parameters.
model=roberta-base
language=en
python finetuning.py --training_languages ${language} --model ${model} --id ${SLURM_ARRAY_TASK_ID}
python extract_lrp_relevance.py --model ${model} --lang ${language} --id ${SLURM_ARRAY_TASK_ID} --case gi
python extract_lrp_relevance.py --model ${model} --lang ${language} --id ${SLURM_ARRAY_TASK_ID} --case lrp
python extract_attention.py --modelname ${model} --lang ${language} --id ${SLURM_ARRAY_TASK_ID}
Total reading times can be extracted by running this script The following parameters need to be set inline depending on the analysis:
MECO = False
WORKERS = "mturk_only" # "mturk_only", "volunteer_only" or None (= "all")
FILTER_QUALITY = False
threshold = None
FILTER_CORRECT_ANSWER = False
FILTER_VISION = False
vision = 'normal' #"glasses", we only have this information for the data recorded at KU
TBC