AmberYifan commited on
Commit
6448c7b
·
verified ·
1 Parent(s): ce8977e

End of training

Browse files
Files changed (5) hide show
  1. README.md +2 -1
  2. all_results.json +8 -0
  3. train_results.json +8 -0
  4. trainer_state.json +127 -0
  5. training_loss.png +0 -0
README.md CHANGED
@@ -4,6 +4,7 @@ license: apache-2.0
4
  base_model: AmberYifan/qwen3-4b-thinking-full-pretrain-mix-low-tweet-1m-en
5
  tags:
6
  - llama-factory
 
7
  - generated_from_trainer
8
  model-index:
9
  - name: qwen3-4b-thinking-full-pretrain-mix-low-tweet-1m-en-sft
@@ -15,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # qwen3-4b-thinking-full-pretrain-mix-low-tweet-1m-en-sft
17
 
18
- This model is a fine-tuned version of [AmberYifan/qwen3-4b-thinking-full-pretrain-mix-low-tweet-1m-en](https://huggingface.co/AmberYifan/qwen3-4b-thinking-full-pretrain-mix-low-tweet-1m-en) on an unknown dataset.
19
 
20
  ## Model description
21
 
 
4
  base_model: AmberYifan/qwen3-4b-thinking-full-pretrain-mix-low-tweet-1m-en
5
  tags:
6
  - llama-factory
7
+ - full
8
  - generated_from_trainer
9
  model-index:
10
  - name: qwen3-4b-thinking-full-pretrain-mix-low-tweet-1m-en-sft
 
16
 
17
  # qwen3-4b-thinking-full-pretrain-mix-low-tweet-1m-en-sft
18
 
19
+ This model is a fine-tuned version of [AmberYifan/qwen3-4b-thinking-full-pretrain-mix-low-tweet-1m-en](https://huggingface.co/AmberYifan/qwen3-4b-thinking-full-pretrain-mix-low-tweet-1m-en) on the alpaca_en dataset.
20
 
21
  ## Model description
22
 
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "total_flos": 3713612709888.0,
4
+ "train_loss": 1.020559032758077,
5
+ "train_runtime": 200.6016,
6
+ "train_samples_per_second": 74.775,
7
+ "train_steps_per_second": 0.598
8
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "total_flos": 3713612709888.0,
4
+ "train_loss": 1.020559032758077,
5
+ "train_runtime": 200.6016,
6
+ "train_samples_per_second": 74.775,
7
+ "train_steps_per_second": 0.598
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,127 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 120,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.25316455696202533,
14
+ "grad_norm": 11.874071671786021,
15
+ "learning_rate": 7.500000000000001e-06,
16
+ "loss": 1.9812,
17
+ "step": 10
18
+ },
19
+ {
20
+ "epoch": 0.5063291139240507,
21
+ "grad_norm": 1.980111719643074,
22
+ "learning_rate": 9.896703108827758e-06,
23
+ "loss": 1.353,
24
+ "step": 20
25
+ },
26
+ {
27
+ "epoch": 0.759493670886076,
28
+ "grad_norm": 1.8905454650238032,
29
+ "learning_rate": 9.401006955900555e-06,
30
+ "loss": 1.2479,
31
+ "step": 30
32
+ },
33
+ {
34
+ "epoch": 1.0,
35
+ "grad_norm": 1.8729820561469692,
36
+ "learning_rate": 8.535533905932739e-06,
37
+ "loss": 1.2144,
38
+ "step": 40
39
+ },
40
+ {
41
+ "epoch": 1.2531645569620253,
42
+ "grad_norm": 1.7342992975506955,
43
+ "learning_rate": 7.373001848738203e-06,
44
+ "loss": 0.9445,
45
+ "step": 50
46
+ },
47
+ {
48
+ "epoch": 1.5063291139240507,
49
+ "grad_norm": 1.8683455108997369,
50
+ "learning_rate": 6.011087861660191e-06,
51
+ "loss": 0.9147,
52
+ "step": 60
53
+ },
54
+ {
55
+ "epoch": 1.759493670886076,
56
+ "grad_norm": 1.9494176703942474,
57
+ "learning_rate": 4.564221286261709e-06,
58
+ "loss": 0.9182,
59
+ "step": 70
60
+ },
61
+ {
62
+ "epoch": 2.0,
63
+ "grad_norm": 2.7049715031754444,
64
+ "learning_rate": 3.1539692634365788e-06,
65
+ "loss": 0.9088,
66
+ "step": 80
67
+ },
68
+ {
69
+ "epoch": 2.2531645569620253,
70
+ "grad_norm": 1.7264154986690055,
71
+ "learning_rate": 1.8988225436587005e-06,
72
+ "loss": 0.7228,
73
+ "step": 90
74
+ },
75
+ {
76
+ "epoch": 2.5063291139240507,
77
+ "grad_norm": 1.6933940779071752,
78
+ "learning_rate": 9.042397785550405e-07,
79
+ "loss": 0.6609,
80
+ "step": 100
81
+ },
82
+ {
83
+ "epoch": 2.759493670886076,
84
+ "grad_norm": 1.7202795598604845,
85
+ "learning_rate": 2.5378678213483057e-07,
86
+ "loss": 0.7076,
87
+ "step": 110
88
+ },
89
+ {
90
+ "epoch": 3.0,
91
+ "grad_norm": 1.7296483993190936,
92
+ "learning_rate": 2.1152495889970035e-09,
93
+ "loss": 0.6727,
94
+ "step": 120
95
+ },
96
+ {
97
+ "epoch": 3.0,
98
+ "step": 120,
99
+ "total_flos": 3713612709888.0,
100
+ "train_loss": 1.020559032758077,
101
+ "train_runtime": 200.6016,
102
+ "train_samples_per_second": 74.775,
103
+ "train_steps_per_second": 0.598
104
+ }
105
+ ],
106
+ "logging_steps": 10,
107
+ "max_steps": 120,
108
+ "num_input_tokens_seen": 0,
109
+ "num_train_epochs": 3,
110
+ "save_steps": 500,
111
+ "stateful_callbacks": {
112
+ "TrainerControl": {
113
+ "args": {
114
+ "should_epoch_stop": false,
115
+ "should_evaluate": false,
116
+ "should_log": false,
117
+ "should_save": true,
118
+ "should_training_stop": true
119
+ },
120
+ "attributes": {}
121
+ }
122
+ },
123
+ "total_flos": 3713612709888.0,
124
+ "train_batch_size": 8,
125
+ "trial_name": null,
126
+ "trial_params": null
127
+ }
training_loss.png ADDED