-
Notifications
You must be signed in to change notification settings - Fork 4
/
Copy pathpile-tiny.yaml
147 lines (130 loc) · 2.87 KB
/
pile-tiny.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
run_name: <run_name>
seed: 6198
dry_run: false
#wandb:
# name: ${run_name}
# log_interval: 100
model:
d_model: 512
n_heads: 8
n_layers: 8
mlp_ratio: 8
alibi: true
alibi_bias_max: 8.0
flash_attention: false
attention_dropout: 0.0
attention_layer_norm: false
multi_query_attention: true
block_type: sequential
layer_norm_type: low_precision # if not compiling, use 'low_precision'
activation_type: gelu
residual_dropout: 0.0
embedding_dropout: 0.0
max_sequence_length: 1024
include_bias: true
vocab_size: 50277
embedding_size: 50304
eos_token_id: 50276
pad_token_id: 50276
init_device: meta
init_std: 0.02
#compile:
#mode: default
optimizer:
name: adamw
learning_rate: 1.0e-3
weight_decay: 0.01
betas:
- 0.9
- 0.95
scheduler:
name: cosine_with_warmup
t_warmup: 4000
alpha_f: 0.1
data:
paths: ${path.glob:<train_dir>/*.npy}
pad_direction: right
num_workers: 2
drop_last: true
pin_memory: true
prefetch_factor: 4 # bump to 16 if on LUMI
persistent_workers: true
timeout: 0
tokenizer:
identifier: EleutherAI/gpt-neox-20b
truncate_direction: right
save_folder: <save_folder>/${run_name}
save_overwrite: true
# Sharded checkpoints (best for restarts)
save_interval: 10000000000
save_num_checkpoints_to_keep: 2
# Unsharded checkpoints (for final storage)
save_interval_unsharded: 1000
save_num_unsharded_checkpoints_to_keep: 2
load_path: null
max_duration: 10000 # 2T tokens
global_train_batch_size: 128
device_train_microbatch_size: 8
precision: amp_bf16
max_grad_norm: 1.0
speed_monitor:
window_size: 20
console_log_interval: 100
matformer_factor: 1
eval_interval: 100
eval_subset_num_batches: 10
device_eval_batch_size: ${device_train_microbatch_size}
evaluators:
##########################
# Perplexity evaluations #
##########################
- label: pile-validation
subset_num_batches: 10
data:
paths: [<Validation paths>]
num_workers: 2
drop_last: true
pin_memory: true
persistent_workers: true
prefetch_factor: 4
##########################
# Downstream evaluations #
##########################
# - label: piqa
# type: downstream
#
# - label: hellaswag
# type: downstream
#
# - label: winogrande
# type: downstream
#
# - label: openbook_qa
# type: downstream
#
# # - label: boolq # requires implemention of the pmi_dc matrix
# # type: downstream
# #
# - label: sciq
# type: downstream
#
# - label: arc_easy
# type: downstream
#
# # - label: arc_challenge # requires implemention of the pmi_dc matrix
# # type: downstream
# #
# - label: copa
# type: downstream
#
# - label: rte
# type: downstream
#
# - label: commitment_bank
# type: downstream
#
# - label: mrpc
# type: downstream
#
# - label: sst2
# type: downstream