Skip to content
/ OFA-X Public

This repository contains the code for the publication "Harnessing the Power of Multi-Task Pretraining for Ground-Truth Level Natural Language Explanations"

License

Notifications You must be signed in to change notification settings

ofa-x/OFA-X

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Harnessing the Power of Multi-Task Pretraining for Ground-Truth Level Natural Language Explanations

Demo page coming soon!

This repository contains the code for the publication "Harnessing the Power of Multi-Task Pretraining for Ground-Truth Level Natural Language Explanations" by Björn Plüster, Jakob Ambsdorf, Lukas Braach, Jae Hee Lee and Stefan Wermter.

It includes a fork of OFA-Sys/OFA (found in the ./OFA directory) and all necessary code to train OFA on VL-NLE tasks (such as VQA-X, e-SNLI-VE, and VCR) for the e-ViL benchmark.

./training contains the code and configurations for training and evaluating the models. The training README contains more information on how to run the training scripts.

./dataset_preparation contains the code for generating the datasets and where to get all required files. See the dataset preparation README for more information.

The ./survey directory contains all data related to the human evaluation conducted in the paper, with more information in the survey survey README.

If you are using OFA-X in your work, please consider citing:

@article{pluster2022harnessing,
  title={Harnessing the Power of Multi-Task Pretraining for Ground-Truth Level Natural Language Explanations},
  author={Pl{\"u}ster, Bj{\"o}rn and Ambsdorf, Jakob and Braach, Lukas and Lee, Jae Hee and Wermter, Stefan},
  journal={arXiv preprint arXiv:2212.04231},
  year={2022}
}

@inproceedings{wang2022ofa,
  title={Ofa: Unifying architectures, tasks, and modalities through a simple sequence-to-sequence learning framework},
  author={Wang, Peng and Yang, An and Men, Rui and Lin, Junyang and Bai, Shuai and Li, Zhikang and Ma, Jianxin and Zhou, Chang and Zhou, Jingren and Yang, Hongxia},
  booktitle={International Conference on Machine Learning},
  pages={23318--23340},
  year={2022},
  organization={PMLR}
}

Model Weights

Please see the links in the table to download the trained model weights. The base-size model is only available with OFA-pretraining, while we selected the huge-size model depending on BERTScore performance of the Large model.

Training Pretraining Model Weights
VQA-X OFA Base, Large
VQA-X Caption Large, Huge
e-SNLI-VE OFA Base, Large, Huge
e-SNLI-VE Caption Large
VCR OFA Base, Large, Huge
VCR Caption Large
OFA-X_MT (e-ViL-comb.) OFA Large

About

This repository contains the code for the publication "Harnessing the Power of Multi-Task Pretraining for Ground-Truth Level Natural Language Explanations"

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published