Skip to content

Commit

Permalink
Adding llama31 (#2404)
Browse files Browse the repository at this point in the history
Summary:
Runs fine locally on CPU, pipelines will hopefully validate that

Pull Request resolved: #2404

Reviewed By: kit1980

Differential Revision: D60628732

Pulled By: xuzhao9

fbshipit-source-id: 7a5a313e871311f2d1c0eadcf125128ccd407743
  • Loading branch information
MaanavD authored and facebook-github-bot committed Aug 2, 2024
1 parent 5aed9cd commit 86c8ec6
Show file tree
Hide file tree
Showing 4 changed files with 40 additions and 0 deletions.
16 changes: 16 additions & 0 deletions torchbenchmark/canary_models/llama_v31_8b/__init__.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,16 @@
from torchbenchmark.tasks import NLP
from torchbenchmark.util.framework.huggingface.model_factory import HuggingFaceModel, HuggingFaceAuthMixin

class Model(HuggingFaceModel, HuggingFaceAuthMixin):
task = NLP.LANGUAGE_MODELING
DEFAULT_TRAIN_BSIZE = 1
DEFAULT_EVAL_BSIZE = 1
DEEPCOPY = False

def __init__(self, test, device, batch_size=None, extra_args=[]):
HuggingFaceAuthMixin.__init__(self)
super().__init__(name="llama_v31_8b", test=test, device=device, batch_size=batch_size, extra_args=extra_args)


def train(self):
return NotImplementedError("FSDP should implement a training loop")
6 changes: 6 additions & 0 deletions torchbenchmark/canary_models/llama_v31_8b/install.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
import os
from torchbenchmark.util.framework.huggingface.patch_hf import cache_model

if __name__ == "__main__":
model_name = os.path.basename(os.path.dirname(os.path.abspath(__file__)))
cache_model(model_name)
12 changes: 12 additions & 0 deletions torchbenchmark/canary_models/llama_v31_8b/metadata.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
devices:
NVIDIA A100-SXM4-40GB:
eval_batch_size: 1
eval_benchmark: false
eval_deterministic: false
eval_nograd: true
not_implemented:
- device: cpu
- device: cuda
test: train
train_benchmark: false
train_deterministic: false
6 changes: 6 additions & 0 deletions torchbenchmark/util/framework/huggingface/basic_configs.py
Original file line number Diff line number Diff line change
Expand Up @@ -137,6 +137,12 @@
'AutoConfig.from_pretrained("meta-llama/Llama-2-70b-hf")',
"AutoModelForMaskedLM",
),
"llama_v31_8b": (
512,
512,
'AutoConfig.from_pretrained("meta-llama/Meta-Llama-3.1-8B")',
"AutoModelForCausalLM",
),
"codellama": (
512,
512,
Expand Down

0 comments on commit 86c8ec6

Please sign in to comment.