Skip to content

Commit

Permalink
Add lost huggingface.yaml (#2408)
Browse files Browse the repository at this point in the history
Summary:
huggingface.yaml was somehow lost while copying commit from PyTorch, see #2405

Pull Request resolved: #2408

Reviewed By: xuzhao9

Differential Revision: D60686116

Pulled By: kit1980

fbshipit-source-id: 3a170084e82676b7c05a4af049e26ef2aa98b14b
  • Loading branch information
kit1980 authored and facebook-github-bot committed Aug 2, 2024
1 parent 86c8ec6 commit ae2c9e8
Showing 1 changed file with 110 additions and 0 deletions.
110 changes: 110 additions & 0 deletions userbenchmark/dynamo/dynamobench/huggingface.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,110 @@
skip:
all:
# Difficult to setup accuracy test because .eval() not supported
- Reformer
# Fails deepcopy
- BlenderbotForConditionalGeneration
- GPTNeoForCausalLM
- GPTNeoForSequenceClassification
# Fails with even batch size = 1
- GPTJForCausalLM
- GPTJForQuestionAnswering

device:
cpu:
# OOMs
- OPTForCausalLM

control_flow:
- AllenaiLongformerBase

batch_size:
# TODO - Fails even after fake tensors
divisors:
AlbertForMaskedLM: 2
AlbertForQuestionAnswering: 2
AllenaiLongformerBase: 2
BartForCausalLM: 2
BartForConditionalGeneration: 2
BertForMaskedLM: 2
BertForQuestionAnswering: 2
BlenderbotForCausalLM: 8
# BlenderbotForConditionalGeneration : 16
BlenderbotSmallForCausalLM: 4
BlenderbotSmallForConditionalGeneration: 2
CamemBert: 2
DebertaForMaskedLM: 4
DebertaForQuestionAnswering: 2
DebertaV2ForMaskedLM: 4
DebertaV2ForQuestionAnswering: 8
DistilBertForMaskedLM: 2
DistilBertForQuestionAnswering: 2
DistillGPT2: 2
ElectraForCausalLM: 2
ElectraForQuestionAnswering: 2
GPT2ForSequenceClassification: 2
# GPTJForCausalLM : 2
# GPTJForQuestionAnswering : 2
# GPTNeoForCausalLM : 32
# GPTNeoForSequenceClassification : 2
GoogleFnet: 2
LayoutLMForMaskedLM: 2
LayoutLMForSequenceClassification: 2
M2M100ForConditionalGeneration: 4
MBartForCausalLM: 2
MBartForConditionalGeneration: 2
MT5ForConditionalGeneration: 2
MegatronBertForCausalLM: 4
MegatronBertForQuestionAnswering: 2
MobileBertForMaskedLM: 2
MobileBertForQuestionAnswering: 2
OPTForCausalLM: 2
PLBartForCausalLM: 2
PLBartForConditionalGeneration: 2
PegasusForCausalLM: 4
PegasusForConditionalGeneration: 2
RobertaForCausalLM: 2
RobertaForQuestionAnswering: 2
Speech2Text2ForCausalLM: 4
T5ForConditionalGeneration: 2
T5Small: 2
TrOCRForCausalLM: 2
XGLMForCausalLM: 4
XLNetLMHeadModel: 2


tolerance:
higher_training:
- MT5ForConditionalGeneration
# AlbertForQuestionAnswering fails in CI GCP A100 but error does not seem
# harmful.
- AlbertForQuestionAnswering

higher_max_autotune_training:
# DebertaForQuestionAnswering needs higher tolerance in Max-Autotune mode
- DebertaForQuestionAnswering

higher_inference:
- GPT2ForSequenceClassification
- RobertaForQuestionAnswering

higher_inference_cpu:
- LayoutLMForSequenceClassification

cosine: []


accuracy:
skip:
large_models:
# Models too large to have eager, dynamo and fp64_numbers simultaneously
# even for 40 GB machine.
- DebertaV2ForMaskedLM
- BlenderbotForCausalLM

only_inference:
# Fails with dynamo for train mode
- M2M100ForConditionalGeneration

only_fp32:
- GoogleFnet

0 comments on commit ae2c9e8

Please sign in to comment.