D-eon commited on
Commit
2828aa5
·
verified ·
1 Parent(s): f829de8

Upload folder using huggingface_hub

Browse files
model/MolmoE-1B-0924-10111756-26000-A1/config.yaml ADDED
@@ -0,0 +1,343 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ run_name: multitask_train
2
+ seed: 6198
3
+ epoch: null
4
+ dry_run: false
5
+ model:
6
+ d_model: 2048
7
+ n_heads: 16
8
+ n_kv_heads: null
9
+ qkv_bias: false
10
+ clip_qkv: null
11
+ n_layers: 16
12
+ mlp_ratio: 1
13
+ mlp_hidden_size: null
14
+ activation_type: swiglu
15
+ block_type: moe
16
+ block_group_size: 1
17
+ rope: true
18
+ rope_full_precision: true
19
+ rope_theta: 10000.0
20
+ vision_backbone:
21
+ image_model_type: openai
22
+ image_default_input_size:
23
+ - 336
24
+ - 336
25
+ image_patch_size: 14
26
+ image_pos_patch_size: 14
27
+ image_emb_dim: 1024
28
+ image_num_heads: 16
29
+ image_num_key_value_heads: 16
30
+ image_num_layers: 23
31
+ image_head_dim: 64
32
+ image_mlp_dim: 4096
33
+ image_mlp_activations: quick_gelu
34
+ image_dropout_rate: 0.0
35
+ image_num_pos: 577
36
+ image_norm_eps: 1.0e-05
37
+ attention_dropout: 0.0
38
+ residual_dropout: 0.0
39
+ initializer_range: 0.02
40
+ fsdp_wrap: false
41
+ resize_mode: default
42
+ vit_load_path: /mnt/data/zhangkaidong/lizihang/weka/oe-training-default/mm-olmo/pretrained_image_encoders/vit-l-14-336.pt
43
+ llm_load_path: /mnt/data/zhangkaidong/lizihang/weka/oe-training-default/mm-olmo/olmoe/model.safetensors
44
+ low_cpu_fsdp: true
45
+ attention_type: sdpa
46
+ float32_attention: true
47
+ attention_dropout: 0.0
48
+ attention_layer_norm: true
49
+ residual_dropout: 0.1
50
+ response_residual_dropout: 0.0
51
+ embedding_dropout: 0.0
52
+ layer_norm_type: rms
53
+ layer_norm_with_affine: true
54
+ layer_norm_eps: 1.0e-05
55
+ attention_layer_norm_with_affine: true
56
+ max_sequence_length: 4096
57
+ max_position_embeddings: 32768
58
+ include_bias: false
59
+ bias_for_layer_norm: false
60
+ scale_logits: false
61
+ vocab_size: 50280
62
+ embedding_size: 50304
63
+ additional_vocab_size: 128
64
+ new_embedding_init_range: 0.02
65
+ weight_tying: false
66
+ init_device: meta
67
+ init_fn: normal
68
+ init_std: 0.02
69
+ init_cutoff_factor: 3.0
70
+ norm_after: false
71
+ precision: amp_bf16
72
+ max_crops: 12
73
+ crop_mode: overlap-and-resize-c2
74
+ use_col_tokens: true
75
+ prompt_type: uber_model
76
+ system_prompt_kind: demo_or_style
77
+ message_formatting: role
78
+ always_start_with_space: true
79
+ multi_annotation_weighting: root_subsegments
80
+ default_inference_len: 128
81
+ overlap_margins:
82
+ - 4
83
+ - 4
84
+ pad_value: 0.0
85
+ image_padding_embed: pad_and_partial_pad
86
+ fix_image_padding: true
87
+ vit_layers:
88
+ - -2
89
+ - -9
90
+ image_pooling_h: 2
91
+ image_pooling_w: 2
92
+ image_pooling_2d: attention_meanq
93
+ image_projector: mlp
94
+ image_feature_dropout: 0.0
95
+ initializer_range: 0.02
96
+ normalize_input_embeds: false
97
+ use_position_ids: true
98
+ head_dim: null
99
+ action_tokenizer:
100
+ identifier: physical-intelligence/fast
101
+ tokenizer_dir: null
102
+ action_dim: 7
103
+ horizon: 8
104
+ tokenizer:
105
+ identifier: allenai/OLMoE-1B-7B-0924
106
+ tokenizer_dir: /mnt/data/zhangkaidong/lizihang/weka/oe-training-default/mm-olmo/tokenizer
107
+ pad_tokenizer: false
108
+ moe_num_experts: 64
109
+ moe_top_k: 8
110
+ moe_mlp_impl: sparse
111
+ moe_log_expert_assignment: false
112
+ moe_shared_expert: false
113
+ moe_lbl_in_fp32: false
114
+ moe_interleave: false
115
+ moe_loss_weight: 0.0
116
+ moe_zloss_weight: 0.0
117
+ moe_dropless: true
118
+ moe_capacity_factor: 1.25
119
+ action_head: l1_regression
120
+ num_diffusion_steps: 1000
121
+ num_diffusion_inference_steps: 30
122
+ use_proprio: false
123
+ action_head_dit_hidden_size: 1024
124
+ action_head_dit_depth: 14
125
+ action_head_dit_num_heads: 16
126
+ llm_causal_attention: false
127
+ action_use_left_eef: false
128
+ action_use_mobile_base: false
129
+ allow_resume: true
130
+ ft_llm: true
131
+ ft_vit: true
132
+ ft_connector: true
133
+ ft_embedding: lm_head
134
+ lora: false
135
+ use_lora: false
136
+ lora_rank: 32
137
+ lora_llm: false
138
+ lora_vit: false
139
+ lora_connector: false
140
+ optimizer:
141
+ name: adamw
142
+ learning_rate: 0.0001
143
+ weight_decay: 0.01
144
+ betas:
145
+ - 0.9
146
+ - 0.95
147
+ eps: 1.0e-05
148
+ connector_learning_rate: 0.0002
149
+ vit_learning_rate: 6.0e-06
150
+ llm_learning_rate: 2.0e-05
151
+ connector_weight_decay: 0.0
152
+ vit_weight_decay: 0.0
153
+ llm_weight_decay: 0.0
154
+ connector_betas:
155
+ - 0.9
156
+ - 0.95
157
+ vit_betas:
158
+ - 0.9
159
+ - 0.95
160
+ llm_betas:
161
+ - 0.9
162
+ - 0.95
163
+ connector_eps: 1.0e-06
164
+ vit_eps: 1.0e-06
165
+ llm_eps: 1.0e-06
166
+ metrics_log_interval: 20
167
+ scheduler:
168
+ name: multimodal
169
+ units: steps
170
+ t_warmup: 100
171
+ t_max: null
172
+ alpha_f: 0.1
173
+ connector_t_warmup: 200
174
+ vit_t_warmup: 500
175
+ llm_t_warmup: 500
176
+ grad_clip_warmup_steps: null
177
+ grad_clip_warmup_factor: null
178
+ warmup_min_lr: 0.0
179
+ data:
180
+ dataset: null
181
+ mixture: null
182
+ root_size_mixture:
183
+ - rate: 0.125
184
+ mixture:
185
+ sr_planning: 500000.0
186
+ robovqa: 500000.0
187
+ - rate: 0.125
188
+ mixture:
189
+ pixmo_ask_model_anything: null
190
+ pixmo_cap: null
191
+ pixmo_points: null
192
+ pixmo_count: null
193
+ blip_laion_cc: null
194
+ - rate: 0.125
195
+ mixture:
196
+ sr_affordance: null
197
+ - rate: 0.125
198
+ mixture:
199
+ sr_trajectory: null
200
+ - rate: 0.5
201
+ mixture:
202
+ oxe_magic_soup_plus_minus_A1: null
203
+ split: train
204
+ seed: 50189
205
+ shuffle_messages: true
206
+ pad: to_max
207
+ sequence_length: 2304
208
+ shuffle: true
209
+ for_inference: false
210
+ multi_modal: torch
211
+ num_workers: 0
212
+ drop_last: true
213
+ pin_memory: true
214
+ prefetch_factor: null
215
+ persistent_workers: false
216
+ timeout: 0
217
+ rlds_dataset_name: ''
218
+ rlds_data_root_dir: null
219
+ use_wrist_image: false
220
+ use_proprio: false
221
+ restore_dataloader: true
222
+ fast_forward_batches: null
223
+ evaluators: []
224
+ eval_interval: 2000
225
+ inf_eval_interval: 2000
226
+ inf_evaluators:
227
+ - label: robovqa
228
+ data:
229
+ dataset: robovqa
230
+ mixture: null
231
+ root_size_mixture: null
232
+ split: validation
233
+ seed: null
234
+ shuffle_messages: true
235
+ pad: to_max
236
+ sequence_length: 1792
237
+ shuffle: true
238
+ for_inference: true
239
+ multi_modal: torch
240
+ num_workers: 0
241
+ drop_last: true
242
+ pin_memory: true
243
+ prefetch_factor: null
244
+ persistent_workers: true
245
+ timeout: 0
246
+ rlds_dataset_name: ''
247
+ rlds_data_root_dir: null
248
+ use_wrist_image: false
249
+ use_proprio: false
250
+ device_eval_batch_size: null
251
+ subset_num_batches: null
252
+ max_examples: 2048
253
+ max_new_tokens: 128
254
+ mm_evaluator:
255
+ n_to_log: 0
256
+ num_wandb_examples: 32
257
+ save_predictions: null
258
+ save_tokens: false
259
+ vqa_eval: robovqa_score
260
+ multi_threshold_box_eval: false
261
+ coordinate_eval: false
262
+ pointing_eval: false
263
+ count_eval: false
264
+ point_count_eval: false
265
+ trajectory_eval: false
266
+ android_eval: false
267
+ clock_eval: false
268
+ clock_bench_eval: false
269
+ math_vista_eval: false
270
+ action_eval: false
271
+ save_dir: null
272
+ save_to_checkpoint_dir: false
273
+ eval_name: null
274
+ skip_if_metrics_cached: true
275
+ save_folder: ./model/checkpoints/10111756
276
+ remote_save_folder: null
277
+ canceled_check_interval: 50
278
+ save_interval: 500
279
+ save_interval_unsharded: 32000
280
+ save_interval_ephemeral: null
281
+ save_interval_action_head: null
282
+ save_num_checkpoints_to_keep: 1
283
+ save_num_unsharded_checkpoints_to_keep: -1
284
+ save_num_action_head_checkpoints_to_keep: -1
285
+ save_overwrite: false
286
+ force_save_unsharded: false
287
+ no_pre_train_checkpoint: true
288
+ initial_model_checkpoint: model/MolmoE-1B-0924-10082005-16000-A1
289
+ load_model_config: null
290
+ load_path: null
291
+ load_path_sharded_checkpointer: null
292
+ reset_optimizer_state: false
293
+ reset_trainer_state: false
294
+ save_dataloader_state: false
295
+ reset_dataloader_state: false
296
+ sharded_checkpointer: torch_legacy
297
+ max_duration: 32000
298
+ global_train_batch_size: 256
299
+ device_train_batch_size: 64
300
+ device_train_microbatch_size: 8
301
+ device_eval_batch_size: 64
302
+ eval_subset_num_batches: 8
303
+ eval_on_load: false
304
+ device_inf_eval_batch_size: 4
305
+ inf_eval_subset_num_batches: -1
306
+ device_train_grad_accum: 8
307
+ max_grad_norm: 1.0
308
+ multi_component_grad_norm: true
309
+ batch_divisor: global_batch
310
+ max_grad_norm_ratio: null
311
+ precision: amp_bf16
312
+ wandb:
313
+ project: molmo_training
314
+ entity: ''
315
+ group: null
316
+ name: MolmoE-1B-0924-A1-pretrain
317
+ tags:
318
+ - watching
319
+ log_artifacts: false
320
+ rank_zero_only: true
321
+ log_interval: 20
322
+ speed_monitor:
323
+ window_size: 20
324
+ gpu_flops_available: null
325
+ console_log_interval: 1
326
+ gen1_gc_interval: 1
327
+ compile: null
328
+ fsdp:
329
+ use_orig_params: true
330
+ sharding_strategy: FULL_SHARD
331
+ wrapping_strategy: by_block_and_size
332
+ precision: float
333
+ hybrid_sharding_num_model_replicas: null
334
+ softmax_auxiliary_loss: true
335
+ softmax_auxiliary_loss_scale: 0.0001
336
+ time_limit: null
337
+ extra_steps_after_cancel: 10
338
+ python_profiling: false
339
+ torch_profiling: false
340
+ stop_at: 32000
341
+ stop_after: null
342
+ activation_checkpointing: whole_layer
343
+ fused_loss: null
model/MolmoE-1B-0924-10111756-26000-A1/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c669be25183a82808f45d383d5009c73481cb867961918fd08ba2bc6cd5961fb
3
+ size 28887690598