ygaci commited on
Commit
4ee3dfb
·
verified ·
1 Parent(s): 16c168c

Training in progress, step 114, checkpoint

Browse files
last-checkpoint/README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit
3
  library_name: peft
4
  ---
5
 
 
1
  ---
2
+ base_model: unsloth/llama-3.2-3b-instruct-bnb-4bit
3
  library_name: peft
4
  ---
5
 
last-checkpoint/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit",
5
  "bias": "none",
6
  "eva_config": null,
7
  "exclude_modules": null,
@@ -25,11 +25,11 @@
25
  "target_modules": [
26
  "gate_proj",
27
  "down_proj",
28
- "o_proj",
29
- "v_proj",
30
  "q_proj",
31
  "k_proj",
32
- "up_proj"
 
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "unsloth/llama-3.2-3b-instruct-bnb-4bit",
5
  "bias": "none",
6
  "eva_config": null,
7
  "exclude_modules": null,
 
25
  "target_modules": [
26
  "gate_proj",
27
  "down_proj",
28
+ "up_proj",
 
29
  "q_proj",
30
  "k_proj",
31
+ "v_proj",
32
+ "o_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:674201cda7990812ad89a0ac22c31df0b82d264ffcdd8a5bdb1dd62baf293d81
3
  size 389074464
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6aec35e09ff7cf30637b3f962aa52b1697faf4ad3cca422cec1820c7f5b65cd
3
  size 389074464
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:93975c5b076ce7d294a4f8acb8ef330cf5533dda15b666fec59a388f8c4aa1a9
3
- size 198011700
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0d3fd82d9118f258b6b2d10b83e40e4c1cce05da4ab9333b5f51df337b36958
3
+ size 198011252
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9196a1e708bf24d6abba41cce3f8558820acc3e50f9394c5955e29eb41ffea3d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:386fcc8cc1089aade9450d86fb239ea3483f455fd2d78d8378645feecfec9d69
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:336ad624cfb7adf5821759ee06496a34d9d2a658dd99d3dbdda29b76da15094d
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd934acdab29b3dbe7dec7ebf9aa0c7d670bf9c37b6ec1d10cfdea1e691d24f3
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,124 +1,26 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.0,
5
  "eval_steps": 500,
6
- "global_step": 1569,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.19120458891013384,
13
- "grad_norm": 0.3175047039985657,
14
- "learning_rate": 4.951737451737452e-05,
15
- "loss": 1.0677,
16
- "step": 100
17
- },
18
- {
19
- "epoch": 0.3824091778202677,
20
- "grad_norm": 0.3452897071838379,
21
- "learning_rate": 4.8552123552123555e-05,
22
- "loss": 0.9592,
23
- "step": 200
24
- },
25
- {
26
- "epoch": 0.5736137667304015,
27
- "grad_norm": 0.5109040141105652,
28
- "learning_rate": 4.758687258687259e-05,
29
- "loss": 0.9311,
30
- "step": 300
31
- },
32
- {
33
- "epoch": 0.7648183556405354,
34
- "grad_norm": 0.5286597609519958,
35
- "learning_rate": 4.662162162162162e-05,
36
- "loss": 0.8667,
37
- "step": 400
38
- },
39
- {
40
- "epoch": 0.9560229445506692,
41
- "grad_norm": 0.40717196464538574,
42
- "learning_rate": 4.565637065637066e-05,
43
- "loss": 0.8142,
44
- "step": 500
45
- },
46
- {
47
- "epoch": 1.147227533460803,
48
- "grad_norm": 0.5934731960296631,
49
- "learning_rate": 4.4691119691119696e-05,
50
- "loss": 0.7658,
51
- "step": 600
52
- },
53
- {
54
- "epoch": 1.338432122370937,
55
- "grad_norm": 0.5729045271873474,
56
- "learning_rate": 4.3725868725868726e-05,
57
- "loss": 0.7155,
58
- "step": 700
59
- },
60
- {
61
- "epoch": 1.5296367112810707,
62
- "grad_norm": 0.8239750862121582,
63
- "learning_rate": 4.276061776061776e-05,
64
- "loss": 0.7286,
65
- "step": 800
66
- },
67
- {
68
- "epoch": 1.7208413001912046,
69
- "grad_norm": 0.7240690588951111,
70
- "learning_rate": 4.17953667953668e-05,
71
- "loss": 0.6913,
72
- "step": 900
73
- },
74
- {
75
- "epoch": 1.9120458891013383,
76
- "grad_norm": 0.6800572872161865,
77
- "learning_rate": 4.083011583011583e-05,
78
- "loss": 0.7385,
79
- "step": 1000
80
- },
81
- {
82
- "epoch": 2.1032504780114722,
83
- "grad_norm": 0.9010747671127319,
84
- "learning_rate": 3.986486486486487e-05,
85
- "loss": 0.6386,
86
- "step": 1100
87
- },
88
- {
89
- "epoch": 2.294455066921606,
90
- "grad_norm": 1.2023605108261108,
91
- "learning_rate": 3.8899613899613905e-05,
92
- "loss": 0.5685,
93
- "step": 1200
94
- },
95
- {
96
- "epoch": 2.48565965583174,
97
- "grad_norm": 1.032483458518982,
98
- "learning_rate": 3.7934362934362935e-05,
99
- "loss": 0.5601,
100
- "step": 1300
101
- },
102
- {
103
- "epoch": 2.676864244741874,
104
- "grad_norm": 0.979926347732544,
105
- "learning_rate": 3.696911196911197e-05,
106
- "loss": 0.6145,
107
- "step": 1400
108
- },
109
- {
110
- "epoch": 2.8680688336520075,
111
- "grad_norm": 0.6987595558166504,
112
- "learning_rate": 3.6003861003861e-05,
113
- "loss": 0.5834,
114
- "step": 1500
115
  }
116
  ],
117
- "logging_steps": 100,
118
- "max_steps": 5230,
119
  "num_input_tokens_seen": 0,
120
  "num_train_epochs": 10,
121
- "save_steps": 500,
122
  "stateful_callbacks": {
123
  "TrainerControl": {
124
  "args": {
@@ -131,7 +33,7 @@
131
  "attributes": {}
132
  }
133
  },
134
- "total_flos": 8.55174687846912e+16,
135
  "train_batch_size": 1,
136
  "trial_name": null,
137
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.09982486865148861,
5
  "eval_steps": 500,
6
+ "global_step": 114,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.09982486865148861,
13
+ "grad_norm": 0.259847491979599,
14
+ "learning_rate": 4.9718557607739666e-05,
15
+ "loss": 1.1436,
16
+ "step": 114
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  }
18
  ],
19
+ "logging_steps": 114,
20
+ "max_steps": 11420,
21
  "num_input_tokens_seen": 0,
22
  "num_train_epochs": 10,
23
+ "save_steps": 114,
24
  "stateful_callbacks": {
25
  "TrainerControl": {
26
  "args": {
 
33
  "attributes": {}
34
  }
35
  },
36
+ "total_flos": 6430276097753088.0,
37
  "train_batch_size": 1,
38
  "trial_name": null,
39
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9961853a27f5829fc673e059d81c28ebfd690e6e74dc2ffd482865d556612698
3
  size 5880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e30f0bfd5e4ce7f4926064b814f4f21d1ca2b5d03ca3afd1066872e09df743b
3
  size 5880