sirr commited on
Commit
eeaf65a
1 Parent(s): c7937b5

Training in progress, step 25, checkpoint

Browse files
last-checkpoint/README.md CHANGED
@@ -4,18 +4,6 @@ library_name: peft
4
  ## Training procedure
5
 
6
 
7
- The following `bitsandbytes` quantization config was used during training:
8
- - quant_method: bitsandbytes
9
- - load_in_8bit: False
10
- - load_in_4bit: True
11
- - llm_int8_threshold: 6.0
12
- - llm_int8_skip_modules: None
13
- - llm_int8_enable_fp32_cpu_offload: False
14
- - llm_int8_has_fp16_weight: False
15
- - bnb_4bit_quant_type: nf4
16
- - bnb_4bit_use_double_quant: True
17
- - bnb_4bit_compute_dtype: float16
18
-
19
  The following `bitsandbytes` quantization config was used during training:
20
  - quant_method: bitsandbytes
21
  - load_in_8bit: False
@@ -29,6 +17,5 @@ The following `bitsandbytes` quantization config was used during training:
29
  - bnb_4bit_compute_dtype: float16
30
  ### Framework versions
31
 
32
- - PEFT 0.5.0
33
 
34
  - PEFT 0.5.0
 
4
  ## Training procedure
5
 
6
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  The following `bitsandbytes` quantization config was used during training:
8
  - quant_method: bitsandbytes
9
  - load_in_8bit: False
 
17
  - bnb_4bit_compute_dtype: float16
18
  ### Framework versions
19
 
 
20
 
21
  - PEFT 0.5.0
last-checkpoint/adapter_config.json CHANGED
@@ -8,7 +8,7 @@
8
  "layers_pattern": null,
9
  "layers_to_transform": null,
10
  "lora_alpha": 8,
11
- "lora_dropout": 0.2,
12
  "modules_to_save": null,
13
  "peft_type": "LORA",
14
  "r": 4,
 
8
  "layers_pattern": null,
9
  "layers_to_transform": null,
10
  "lora_alpha": 8,
11
+ "lora_dropout": 0.05,
12
  "modules_to_save": null,
13
  "peft_type": "LORA",
14
  "r": 4,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6614ad059f21b421670cd5e428c91ab88f769eb2a2492105cc8a6983fe3edfa0
3
  size 40036040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a97e01168bb7705aceb33e32e0e4e3d587f3e1163b2c422a18c4e5807a1ccdc
3
  size 40036040
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:901c5fd7af5774f7e3e7f030e0200f5e17298e36cd8fcec8511b2c3751bd4dec
3
  size 20523679
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8851a73de47dbfcfe31de86bb6192dfb9990a4ac4a46f2f7ab1302c1c78989d2
3
  size 20523679
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:30b6b81c68bccbea71620607bbde917982ac0e318eae54d911e6a4cadfae0a72
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3029f8fdae5b5a1515326241e0c15afad3612dc0f653bb9b6b93ae183483db93
3
  size 14575
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a164590e485ad70f68d6dc5c9f98d12a7490944c208738b6125bd2b698b648a9
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01bd1906bbea59f9e9fca9d2fcebc7d848ace4b2b715eed5519ef909ccf2df1c
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,32 +1,32 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.2,
5
- "eval_steps": 100,
6
- "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.2,
13
- "learning_rate": 0.04,
14
- "loss": 10997.5238,
15
- "step": 100
16
  },
17
  {
18
- "epoch": 0.2,
19
- "eval_runtime": 2114.359,
20
- "eval_samples_per_second": 1.892,
21
- "eval_steps_per_second": 0.236,
22
- "step": 100
23
  }
24
  ],
25
- "logging_steps": 100,
26
- "max_steps": 1000,
27
  "num_train_epochs": 2,
28
- "save_steps": 100,
29
- "total_flos": 3.32202164355072e+16,
30
  "trial_name": null,
31
  "trial_params": null
32
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.25,
5
+ "eval_steps": 25,
6
+ "global_step": 25,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.25,
13
+ "learning_rate": 0.0001,
14
+ "loss": 1.6153,
15
+ "step": 25
16
  },
17
  {
18
+ "epoch": 0.25,
19
+ "eval_runtime": 216.6929,
20
+ "eval_samples_per_second": 1.846,
21
+ "eval_steps_per_second": 0.157,
22
+ "step": 25
23
  }
24
  ],
25
+ "logging_steps": 25,
26
+ "max_steps": 200,
27
  "num_train_epochs": 2,
28
+ "save_steps": 25,
29
+ "total_flos": 4152527054438400.0,
30
  "trial_name": null,
31
  "trial_params": null
32
  }
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed3fc0fab185d4e779aef22afa57b457c6d7621ef7637b0afb63fb96731ca99e
3
  size 4091
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c8c27537b3f48a590e7245cfc358a5c006cb5add7444724e14c4fd4a1f89680
3
  size 4091