From 6d0a97de23728850ca0b3aa971e217ce4c0b9b30 Mon Sep 17 00:00:00 2001 From: Jingyey Date: Thu, 7 Mar 2024 22:00:26 -0500 Subject: [PATCH] Update README.md Some description of the two scripts --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 4954926..e176f2a 100644 --- a/README.md +++ b/README.md @@ -1,7 +1,7 @@ # Reversal_Curse This page contains three python notebooks, [Reversal_data_generation](https://github.com/WGLab/Reversal_Curse/blob/main/Reversal_data_generation.ipynb), [llama_union_intersectoin](https://github.com/WGLab/Reversal_Curse/blob/main/llama_union_intersection.ipynb) and [bert_reversal_curse](https://github.com/WGLab/Reversal_Curse/blob/main/bert_reversal_curse.ipynb), for the [paper](https://arxiv.org/abs/2312.03633) *Yang, Jingye, Da Wu, and Kai Wang. "Not All Large Language Models (LLMs) Succumb to the" Reversal Curse": A Comparative Study of Deductive Logical Reasoning in BERT and GPT Models." arXiv preprint arXiv:2312.03633 (2023).* -Specifically, the [Reversal_data_generation](https://github.com/WGLab/Reversal_Curse/blob/main/Reversal_data_generation.ipynb) contains all the code for generating synthetic training and testing data in the paper. The [BERT_reversal](https://github.com/WGLab/Reversal_Curse/blob/main/BERT_reversal.ipynb) contains the code for all the training and evulation on BERT model, whereas [llama_union_intersectoin](https://github.com/WGLab/Reversal_Curse/blob/main/llama_union_intersection.ipynb) contains the code for all the training and evulation on LlaMA model. +Specifically, the [Reversal_data_generation](https://github.com/WGLab/Reversal_Curse/blob/main/Reversal_data_generation.ipynb) contains all the code for generating synthetic training and testing data in the paper. The [bert_reversal_curse](https://github.com/WGLab/Reversal_Curse/blob/main/bert_reversal_curse.ipynb) contains the code for all the training and evaluation of the BERT model, and [llama_union_intersectoin](https://github.com/WGLab/Reversal_Curse/blob/main/llama_union_intersection.ipynb) contains the code for all the training and evulation of LlaMA model. These two scripts help reproduce the main results of the paper (Table 1 - 5). -Furthermore, for those inclined, these code, especially the code for generating synthetic data, can be readily modified and employed for personalized testing and exploration. This flexibility allows interested individuals to tailor the code to their specific requirements and objectives. +Furthermore, for those inclined, these codes, especially the code for generating synthetic data, can be readily modified and employed for personalized testing and exploration. This flexibility allows interested individuals to tailor the code to their specific requirements and objectives.