impossibleexchange commited on
Commit
a651964
1 Parent(s): 02c7332

Upload training_config.yml with huggingface_hub

Browse files
Files changed (1) hide show
  1. training_config.yml +92 -0
training_config.yml ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ _component_: models.lora_mmllama3_8b
3
+ lora_attn_modules:
4
+ - q_proj
5
+ - v_proj
6
+ apply_lora_to_mlp: false
7
+ apply_lora_to_output: false
8
+ lora_rank: 32
9
+ lora_alpha: 64
10
+ perception_tokens: 2
11
+ use_clip: false
12
+ tokenizer:
13
+ _component_: models.a2a_tokenizer
14
+ path: models/tokenizer.model
15
+ checkpointer:
16
+ _component_: torchtune.utils.FullModelMetaCheckpointer
17
+ checkpoint_dir: output_checkpoints/experiment_1
18
+ checkpoint_files:
19
+ - meta_model_0.pt
20
+ adapter_checkpoint: null
21
+ recipe_checkpoint: null
22
+ output_dir: output_checkpoints/experiment_1
23
+ model_type: LLAMA3
24
+ resume_from_checkpoint: false
25
+ interim_checkpoint_steps: 15000
26
+ interim_gen_steps: null
27
+ max_new_tokens: 100
28
+ temperature: 0.8
29
+ top_k: 300
30
+ dataset:
31
+ _component_: ds.EvenBatcher
32
+ buffer_size: 1000
33
+ dataset:
34
+ _component_: ds.RoundRobinDataset
35
+ datasets:
36
+ - _component_: ds.OmegaVideoCaptionDataset
37
+ length: 600000
38
+ - _component_: ds.LlavaInstructDataset
39
+ dataset_path: ds/coco_llava_instruct/output.parquet
40
+ train_on_input: false
41
+ - _component_: ds.LlavaInstructDataset
42
+ dataset_path: ds/vision_flan/output.parquet
43
+ train_on_input: false
44
+ - _component_: ds.CaptionInstructDataset
45
+ dataset_path: ds/sam_llava/output.parquet
46
+ train_on_input: false
47
+ seed: null
48
+ shuffle: true
49
+ batch_size: 6
50
+ optimizer:
51
+ _component_: torch.optim.AdamW
52
+ weight_decay: 0.01
53
+ lr: 5.0e-05
54
+ betas:
55
+ - 0.9
56
+ - 0.999
57
+ lr_scheduler:
58
+ _component_: torchtune.modules.get_cosine_schedule_with_warmup
59
+ num_warmup_steps: 2000
60
+ loss:
61
+ _component_: torch.nn.CrossEntropyLoss
62
+ label_smoothing: 0.1
63
+ grad_clip:
64
+ _component_: torch.nn.utils.clip_grad_norm_
65
+ max_norm: 2.0
66
+ norm_type: 2
67
+ epochs: 6
68
+ max_steps_per_epoch: null
69
+ gradient_accumulation_steps: 16
70
+ compile: false
71
+ output_dir: /tmp/lora_finetune_output
72
+ metric_logger:
73
+ _component_: torchtune.utils.metric_logging.DiskLogger
74
+ log_dir: ${output_dir}
75
+ log_every_n_steps: null
76
+ device: cuda
77
+ dtype: bf16
78
+ enable_activation_checkpointing: false
79
+ profiler:
80
+ _component_: torchtune.utils.profiler
81
+ enabled: false
82
+ inference:
83
+ prompt_template: 'Video:
84
+
85
+ {video}
86
+
87
+ Caption the previous video.'
88
+ max_new_tokens: 300
89
+ temperature: 0.8
90
+ top_k: 300
91
+ quantizer: null
92
+ gradient-accumulation-steps: 32