zuona commited on
Commit
aecbe3e
·
verified ·
1 Parent(s): 2bd846d

End of training

Browse files
Files changed (2) hide show
  1. README.md +3 -1
  2. trainer_state.json +23 -23
README.md CHANGED
@@ -3,6 +3,8 @@ library_name: peft
3
  license: cc-by-nc-4.0
4
  base_model: facebook/musicgen-melody
5
  tags:
 
 
6
  - generated_from_trainer
7
  model-index:
8
  - name: musicgen-melody-lora-baishidaoren
@@ -14,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # musicgen-melody-lora-baishidaoren
16
 
17
- This model is a fine-tuned version of [facebook/musicgen-melody](https://huggingface.co/facebook/musicgen-melody) on an unknown dataset.
18
 
19
  ## Model description
20
 
 
3
  license: cc-by-nc-4.0
4
  base_model: facebook/musicgen-melody
5
  tags:
6
+ - text-to-audio
7
+ - baoshidaoren_music_snippets
8
  - generated_from_trainer
9
  model-index:
10
  - name: musicgen-melody-lora-baishidaoren
 
16
 
17
  # musicgen-melody-lora-baishidaoren
18
 
19
+ This model is a fine-tuned version of [facebook/musicgen-melody](https://huggingface.co/facebook/musicgen-melody) on the zuona/baoshidaoren_music_snippets dataset.
20
 
21
  ## Model description
22
 
trainer_state.json CHANGED
@@ -10,61 +10,61 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.25396825396825395,
13
- "grad_norm": 14.082204818725586,
14
  "learning_rate": 0.00017142857142857143,
15
- "loss": 74.6049,
16
  "step": 2
17
  },
18
  {
19
  "epoch": 0.5079365079365079,
20
- "grad_norm": 21.457565307617188,
21
  "learning_rate": 0.00014285714285714287,
22
- "loss": 69.937,
23
  "step": 4
24
  },
25
  {
26
  "epoch": 0.7619047619047619,
27
- "grad_norm": 21.216873168945312,
28
- "learning_rate": 0.00011428571428571428,
29
- "loss": 63.225,
30
  "step": 6
31
  },
32
  {
33
  "epoch": 1.126984126984127,
34
- "grad_norm": 29.787534713745117,
35
- "learning_rate": 8.571428571428571e-05,
36
- "loss": 78.8573,
37
  "step": 8
38
  },
39
  {
40
  "epoch": 1.380952380952381,
41
- "grad_norm": 19.73782730102539,
42
- "learning_rate": 5.714285714285714e-05,
43
- "loss": 54.8663,
44
  "step": 10
45
  },
46
  {
47
  "epoch": 1.6349206349206349,
48
- "grad_norm": NaN,
49
  "learning_rate": 4.2857142857142856e-05,
50
- "loss": 53.0713,
51
  "step": 12
52
  },
53
  {
54
  "epoch": 1.8888888888888888,
55
- "grad_norm": 14.591973304748535,
56
  "learning_rate": 1.4285714285714285e-05,
57
- "loss": 52.1387,
58
  "step": 14
59
  },
60
  {
61
  "epoch": 1.8888888888888888,
62
  "step": 14,
63
- "total_flos": 64781410181808.0,
64
- "train_loss": 63.81435339791434,
65
- "train_runtime": 494.586,
66
- "train_samples_per_second": 0.51,
67
- "train_steps_per_second": 0.028
68
  }
69
  ],
70
  "logging_steps": 2,
@@ -84,7 +84,7 @@
84
  "attributes": {}
85
  }
86
  },
87
- "total_flos": 64781410181808.0,
88
  "train_batch_size": 2,
89
  "trial_name": null,
90
  "trial_params": null
 
10
  "log_history": [
11
  {
12
  "epoch": 0.25396825396825395,
13
+ "grad_norm": 22.248607635498047,
14
  "learning_rate": 0.00017142857142857143,
15
+ "loss": 72.5258,
16
  "step": 2
17
  },
18
  {
19
  "epoch": 0.5079365079365079,
20
+ "grad_norm": 27.004058837890625,
21
  "learning_rate": 0.00014285714285714287,
22
+ "loss": 66.9012,
23
  "step": 4
24
  },
25
  {
26
  "epoch": 0.7619047619047619,
27
+ "grad_norm": 34.10637664794922,
28
+ "learning_rate": 0.00012857142857142858,
29
+ "loss": 62.4167,
30
  "step": 6
31
  },
32
  {
33
  "epoch": 1.126984126984127,
34
+ "grad_norm": 51.732906341552734,
35
+ "learning_rate": 0.0001,
36
+ "loss": 76.2196,
37
  "step": 8
38
  },
39
  {
40
  "epoch": 1.380952380952381,
41
+ "grad_norm": 32.293487548828125,
42
+ "learning_rate": 7.142857142857143e-05,
43
+ "loss": 52.813,
44
  "step": 10
45
  },
46
  {
47
  "epoch": 1.6349206349206349,
48
+ "grad_norm": 18.748706817626953,
49
  "learning_rate": 4.2857142857142856e-05,
50
+ "loss": 48.6479,
51
  "step": 12
52
  },
53
  {
54
  "epoch": 1.8888888888888888,
55
+ "grad_norm": 21.977031707763672,
56
  "learning_rate": 1.4285714285714285e-05,
57
+ "loss": 49.3026,
58
  "step": 14
59
  },
60
  {
61
  "epoch": 1.8888888888888888,
62
  "step": 14,
63
+ "total_flos": 257111615139720.0,
64
+ "train_loss": 61.260969979422434,
65
+ "train_runtime": 535.6705,
66
+ "train_samples_per_second": 0.47,
67
+ "train_steps_per_second": 0.026
68
  }
69
  ],
70
  "logging_steps": 2,
 
84
  "attributes": {}
85
  }
86
  },
87
+ "total_flos": 257111615139720.0,
88
  "train_batch_size": 2,
89
  "trial_name": null,
90
  "trial_params": null