-
Notifications
You must be signed in to change notification settings - Fork 278
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
llama v2 7b #1775
llama v2 7b #1775
Conversation
… into msaroufim/authsd
Ok weird there's an OOM error here even though this is passing locally on my A10G |
- device: cuda | ||
test: train | ||
- device: cuda | ||
test: example |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I am wondering why cuda example is disabled? Is it because of OOM?
The example test will check whether the output tensor is deterministic across runs, and it will be useful for the accuracy tests.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yep at num_heads = 16 this was the only test that OOM'd because of some cloning - lemme get the error message
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If cloning is an issue, you can add DEEPCOPY = False
to disable it, similar to
DEEPCOPY = False |
Can you please add DEEPCOPY = False
and try again?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Alright trying it now in CI
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ok with deep copy off it still ooms at the full model size
Could you please paste the command and output of running the model locally?
|
e0f091b
to
9fbceda
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for adding llama v2 model!
@msaroufim has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
@msaroufim merged this pull request in 196b3b8. |
No description provided.