-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: experiment with Llama2 (resolves #18) #19
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@cindyli, I found a few things for you to think about changing. However, I do not have the expertise regarding the specifics of training Llama-2 nor testing it , nor the HuggingFace interface. You might want to have someone else with more experience to look that over, but I'm not sure who that would be. Angel?
@klown thanks for the review and great suggestions. All comments are addressed. It's ready for another look. I emailed David and Angel to find out if they could review and provide feedback. Let's see how it goes. |
@agamba as we discussed in the technical meeting yesterday, in the |
All models I downloaded are placed in |
@klown all new comments have been addressed. Ready for another look. Thanks. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@cindyli, there is one last unused variable to get rid of, and then it's good as far as I'm concerned.
jobs/Llama2/finetune/eval_7b_hf.py
Outdated
@@ -46,20 +48,19 @@ def generate_text_with_prompt(prompt, model, tokenizer, temperature=0.7): | |||
# Test with exactly same instructions used in fine-tuning | |||
instruction = "### Instruction: \nConvert the input English sentence to a Bliss sentence.\n\n" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The instruction
variable is not needed anymore, since the instructions_map
is used instead, on the next and subsequent lines.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Great catch, @klown. Now addressed. Thanks.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
The documentation in docs/Llama2FineTuning.md describes experiments with Llama2 models and results. This pull request also includes scripts used in the experiment.