izzcw commited on
Commit
4bba050
·
verified ·
1 Parent(s): 754213b

End of training

Browse files
README.md CHANGED
@@ -4,6 +4,7 @@ license: apache-2.0
4
  base_model: Qwen/Qwen2.5-1.5B-Instruct
5
  tags:
6
  - llama-factory
 
7
  - generated_from_trainer
8
  model-index:
9
  - name: large_crafting_sft_success
@@ -15,9 +16,9 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # large_crafting_sft_success
17
 
18
- This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.3886
21
 
22
  ## Model description
23
 
 
4
  base_model: Qwen/Qwen2.5-1.5B-Instruct
5
  tags:
6
  - llama-factory
7
+ - full
8
  - generated_from_trainer
9
  model-index:
10
  - name: large_crafting_sft_success
 
16
 
17
  # large_crafting_sft_success
18
 
19
+ This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) on the identity and the large_crafting_sft_success datasets.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.3746
22
 
23
  ## Model description
24
 
all_results.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
  "epoch": 0.9936628643852978,
3
- "eval_loss": 0.37712058424949646,
4
- "eval_runtime": 10.3775,
5
- "eval_samples_per_second": 37.292,
6
- "eval_steps_per_second": 4.722,
7
- "total_flos": 93124196827136.0,
8
- "train_loss": 0.5275116664211766,
9
- "train_runtime": 2083.3896,
10
- "train_samples_per_second": 9.088,
11
- "train_steps_per_second": 0.071
12
  }
 
1
  {
2
  "epoch": 0.9936628643852978,
3
+ "eval_loss": 0.3745827078819275,
4
+ "eval_runtime": 6.4057,
5
+ "eval_samples_per_second": 60.415,
6
+ "eval_steps_per_second": 7.649,
7
+ "total_flos": 55493244223488.0,
8
+ "train_loss": 0.5211580756570207,
9
+ "train_runtime": 1269.1418,
10
+ "train_samples_per_second": 14.918,
11
+ "train_steps_per_second": 0.116
12
  }
eval_results.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "epoch": 0.9936628643852978,
3
- "eval_loss": 0.37712058424949646,
4
- "eval_runtime": 10.3775,
5
- "eval_samples_per_second": 37.292,
6
- "eval_steps_per_second": 4.722
7
  }
 
1
  {
2
  "epoch": 0.9936628643852978,
3
+ "eval_loss": 0.3745827078819275,
4
+ "eval_runtime": 6.4057,
5
+ "eval_samples_per_second": 60.415,
6
+ "eval_steps_per_second": 7.649
7
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 0.9936628643852978,
3
- "total_flos": 93124196827136.0,
4
- "train_loss": 0.5275116664211766,
5
- "train_runtime": 2083.3896,
6
- "train_samples_per_second": 9.088,
7
- "train_steps_per_second": 0.071
8
  }
 
1
  {
2
  "epoch": 0.9936628643852978,
3
+ "total_flos": 55493244223488.0,
4
+ "train_loss": 0.5211580756570207,
5
+ "train_runtime": 1269.1418,
6
+ "train_samples_per_second": 14.918,
7
+ "train_steps_per_second": 0.116
8
  }
trainer_state.json CHANGED
@@ -10,126 +10,126 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.06759611322348964,
13
- "grad_norm": 7.767644079038259,
14
  "learning_rate": 6.666666666666667e-06,
15
- "loss": 1.0089,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.13519222644697929,
20
- "grad_norm": 4.34859810909518,
21
  "learning_rate": 9.964639423366442e-06,
22
- "loss": 0.6297,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 0.20278833967046894,
27
- "grad_norm": 4.288987850642454,
28
  "learning_rate": 9.68474862499881e-06,
29
- "loss": 0.5793,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 0.27038445289395857,
34
- "grad_norm": 4.544358861090749,
35
  "learning_rate": 9.140746393556853e-06,
36
- "loss": 0.5668,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 0.33798056611744826,
41
- "grad_norm": 3.502106036516604,
42
  "learning_rate": 8.363301868506264e-06,
43
- "loss": 0.5287,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 0.33798056611744826,
48
- "eval_loss": 0.47872456908226013,
49
- "eval_runtime": 10.4806,
50
- "eval_samples_per_second": 36.926,
51
- "eval_steps_per_second": 4.675,
52
  "step": 50
53
  },
54
  {
55
  "epoch": 0.4055766793409379,
56
- "grad_norm": 3.3511808877782037,
57
  "learning_rate": 7.396244933600285e-06,
58
- "loss": 0.5138,
59
  "step": 60
60
  },
61
  {
62
  "epoch": 0.47317279256442757,
63
- "grad_norm": 3.4962865599588757,
64
  "learning_rate": 6.294095225512604e-06,
65
- "loss": 0.4927,
66
  "step": 70
67
  },
68
  {
69
  "epoch": 0.5407689057879171,
70
- "grad_norm": 3.067612201538514,
71
  "learning_rate": 5.118988487730537e-06,
72
- "loss": 0.4732,
73
  "step": 80
74
  },
75
  {
76
  "epoch": 0.6083650190114068,
77
- "grad_norm": 3.2219742120208332,
78
  "learning_rate": 3.937173552235117e-06,
79
- "loss": 0.4755,
80
  "step": 90
81
  },
82
  {
83
  "epoch": 0.6759611322348965,
84
- "grad_norm": 3.0639384175557667,
85
  "learning_rate": 2.8152774381532033e-06,
86
- "loss": 0.4548,
87
  "step": 100
88
  },
89
  {
90
  "epoch": 0.6759611322348965,
91
- "eval_loss": 0.4024154245853424,
92
- "eval_runtime": 10.4735,
93
- "eval_samples_per_second": 36.95,
94
- "eval_steps_per_second": 4.678,
95
  "step": 100
96
  },
97
  {
98
  "epoch": 0.7435572454583862,
99
- "grad_norm": 2.9634825546836465,
100
  "learning_rate": 1.8165491294045596e-06,
101
- "loss": 0.4643,
102
  "step": 110
103
  },
104
  {
105
  "epoch": 0.8111533586818758,
106
- "grad_norm": 2.6864984928356512,
107
  "learning_rate": 9.972937953781985e-07,
108
- "loss": 0.4229,
109
  "step": 120
110
  },
111
  {
112
  "epoch": 0.8787494719053655,
113
- "grad_norm": 4.377547771161915,
114
  "learning_rate": 4.036984820916723e-07,
115
- "loss": 0.4149,
116
  "step": 130
117
  },
118
  {
119
  "epoch": 0.9463455851288551,
120
- "grad_norm": 2.8491548740185846,
121
  "learning_rate": 6.922823140906754e-08,
122
- "loss": 0.4439,
123
  "step": 140
124
  },
125
  {
126
  "epoch": 0.9936628643852978,
127
  "step": 147,
128
- "total_flos": 93124196827136.0,
129
- "train_loss": 0.5275116664211766,
130
- "train_runtime": 2083.3896,
131
- "train_samples_per_second": 9.088,
132
- "train_steps_per_second": 0.071
133
  }
134
  ],
135
  "logging_steps": 10,
@@ -149,7 +149,7 @@
149
  "attributes": {}
150
  }
151
  },
152
- "total_flos": 93124196827136.0,
153
  "train_batch_size": 1,
154
  "trial_name": null,
155
  "trial_params": null
 
10
  "log_history": [
11
  {
12
  "epoch": 0.06759611322348964,
13
+ "grad_norm": 6.16084466512128,
14
  "learning_rate": 6.666666666666667e-06,
15
+ "loss": 1.0855,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.13519222644697929,
20
+ "grad_norm": 3.497086396174302,
21
  "learning_rate": 9.964639423366442e-06,
22
+ "loss": 0.7122,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 0.20278833967046894,
27
+ "grad_norm": 2.6933889117498127,
28
  "learning_rate": 9.68474862499881e-06,
29
+ "loss": 0.5732,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 0.27038445289395857,
34
+ "grad_norm": 2.6966538346452094,
35
  "learning_rate": 9.140746393556853e-06,
36
+ "loss": 0.5274,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 0.33798056611744826,
41
+ "grad_norm": 2.709346620508466,
42
  "learning_rate": 8.363301868506264e-06,
43
+ "loss": 0.4945,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 0.33798056611744826,
48
+ "eval_loss": 0.4356246590614319,
49
+ "eval_runtime": 6.3988,
50
+ "eval_samples_per_second": 60.48,
51
+ "eval_steps_per_second": 7.658,
52
  "step": 50
53
  },
54
  {
55
  "epoch": 0.4055766793409379,
56
+ "grad_norm": 3.1656760844291267,
57
  "learning_rate": 7.396244933600285e-06,
58
+ "loss": 0.476,
59
  "step": 60
60
  },
61
  {
62
  "epoch": 0.47317279256442757,
63
+ "grad_norm": 2.5563614993729598,
64
  "learning_rate": 6.294095225512604e-06,
65
+ "loss": 0.4591,
66
  "step": 70
67
  },
68
  {
69
  "epoch": 0.5407689057879171,
70
+ "grad_norm": 2.392298112754115,
71
  "learning_rate": 5.118988487730537e-06,
72
+ "loss": 0.44,
73
  "step": 80
74
  },
75
  {
76
  "epoch": 0.6083650190114068,
77
+ "grad_norm": 2.598895314476247,
78
  "learning_rate": 3.937173552235117e-06,
79
+ "loss": 0.4577,
80
  "step": 90
81
  },
82
  {
83
  "epoch": 0.6759611322348965,
84
+ "grad_norm": 2.3301001202203975,
85
  "learning_rate": 2.8152774381532033e-06,
86
+ "loss": 0.4438,
87
  "step": 100
88
  },
89
  {
90
  "epoch": 0.6759611322348965,
91
+ "eval_loss": 0.3885859251022339,
92
+ "eval_runtime": 6.3869,
93
+ "eval_samples_per_second": 60.593,
94
+ "eval_steps_per_second": 7.672,
95
  "step": 100
96
  },
97
  {
98
  "epoch": 0.7435572454583862,
99
+ "grad_norm": 2.41225835615904,
100
  "learning_rate": 1.8165491294045596e-06,
101
+ "loss": 0.44,
102
  "step": 110
103
  },
104
  {
105
  "epoch": 0.8111533586818758,
106
+ "grad_norm": 2.3212619903795653,
107
  "learning_rate": 9.972937953781985e-07,
108
+ "loss": 0.4203,
109
  "step": 120
110
  },
111
  {
112
  "epoch": 0.8787494719053655,
113
+ "grad_norm": 2.3281357946059558,
114
  "learning_rate": 4.036984820916723e-07,
115
+ "loss": 0.4026,
116
  "step": 130
117
  },
118
  {
119
  "epoch": 0.9463455851288551,
120
+ "grad_norm": 2.4364627230218816,
121
  "learning_rate": 6.922823140906754e-08,
122
+ "loss": 0.4334,
123
  "step": 140
124
  },
125
  {
126
  "epoch": 0.9936628643852978,
127
  "step": 147,
128
+ "total_flos": 55493244223488.0,
129
+ "train_loss": 0.5211580756570207,
130
+ "train_runtime": 1269.1418,
131
+ "train_samples_per_second": 14.918,
132
+ "train_steps_per_second": 0.116
133
  }
134
  ],
135
  "logging_steps": 10,
 
149
  "attributes": {}
150
  }
151
  },
152
+ "total_flos": 55493244223488.0,
153
  "train_batch_size": 1,
154
  "trial_name": null,
155
  "trial_params": null
training_eval_loss.png CHANGED
training_loss.png CHANGED