youssefkhalil320 commited on
Commit
5e85ff0
·
verified ·
1 Parent(s): c87b868

Upload folder using huggingface_hub

Browse files
checkpoint-1000/README.md CHANGED
@@ -13,31 +13,31 @@ tags:
13
  - dataset_size:74247
14
  - loss:AnglELoss
15
  widget:
16
- - source_sentence: hand of fatima decanter
17
- sentences:
18
- - matte concealer
19
- - car refresher spray
20
- - bread
21
- - source_sentence: intensive antiwrinkle face cream
22
  sentences:
23
  - bread
24
- - bread
25
  - car refresher spray
26
- - source_sentence: aromatic coffee capsules
 
27
  sentences:
28
- - bread
29
- - car refresher spray
30
- - bread
31
- - source_sentence: side fringe jacket
32
  sentences:
33
- - car refresher spray
34
- - car refresher spray
 
 
 
 
35
  - home scent
36
- - source_sentence: pastelcolored cookies
 
37
  sentences:
38
- - matte concealer
39
- - rice
40
- - bisk
41
  ---
42
 
43
  # all-MiniLM-L6-v6-pair_score
@@ -90,9 +90,9 @@ from sentence_transformers import SentenceTransformer
90
  model = SentenceTransformer("sentence_transformers_model_id")
91
  # Run inference
92
  sentences = [
93
- 'pastelcolored cookies',
94
- 'bisk',
95
- 'rice',
96
  ]
97
  embeddings = model.encode(sentences)
98
  print(embeddings.shape)
@@ -149,7 +149,7 @@ You can finetune this model on your own dataset.
149
  - `per_device_train_batch_size`: 128
150
  - `per_device_eval_batch_size`: 128
151
  - `learning_rate`: 2e-05
152
- - `num_train_epochs`: 2
153
  - `warmup_ratio`: 0.1
154
  - `fp16`: True
155
 
@@ -173,7 +173,7 @@ You can finetune this model on your own dataset.
173
  - `adam_beta2`: 0.999
174
  - `adam_epsilon`: 1e-08
175
  - `max_grad_norm`: 1.0
176
- - `num_train_epochs`: 2
177
  - `max_steps`: -1
178
  - `lr_scheduler_type`: linear
179
  - `lr_scheduler_kwargs`: {}
@@ -274,16 +274,16 @@ You can finetune this model on your own dataset.
274
  ### Training Logs
275
  | Epoch | Step | Training Loss |
276
  |:------:|:----:|:-------------:|
277
- | 0.1721 | 100 | 8.9697 |
278
- | 0.3442 | 200 | 2.6119 |
279
- | 0.5164 | 300 | 0.9351 |
280
- | 0.6885 | 400 | 0.7049 |
281
- | 0.8606 | 500 | 0.394 |
282
- | 1.0327 | 600 | 0.6792 |
283
- | 1.2048 | 700 | 0.715 |
284
- | 1.3769 | 800 | 0.5655 |
285
- | 1.5491 | 900 | 0.5354 |
286
- | 1.7212 | 1000 | 0.6299 |
287
 
288
 
289
  ### Framework Versions
 
13
  - dataset_size:74247
14
  - loss:AnglELoss
15
  widget:
16
+ - source_sentence: paste
 
 
 
 
 
17
  sentences:
18
  - bread
 
19
  - car refresher spray
20
+ - hot drinks mug
21
+ - source_sentence: shoulder strap accessory for bags
22
  sentences:
23
+ - home scent
24
+ - ball
25
+ - one outer pocket cross bag
26
+ - source_sentence: deep fried mozzarella sticks
27
  sentences:
28
+ - one outer pocket cross bag
29
+ - insect killer
30
+ - stir fry vegetables bowl
31
+ - source_sentence: metal flower set tableaus
32
+ sentences:
33
+ - home scent
34
  - home scent
35
+ - hygiene
36
+ - source_sentence: basic choker
37
  sentences:
38
+ - hygiene
39
+ - unisex sweatshirt
40
+ - unisex sweatshirt
41
  ---
42
 
43
  # all-MiniLM-L6-v6-pair_score
 
90
  model = SentenceTransformer("sentence_transformers_model_id")
91
  # Run inference
92
  sentences = [
93
+ 'basic choker',
94
+ 'unisex sweatshirt',
95
+ 'unisex sweatshirt',
96
  ]
97
  embeddings = model.encode(sentences)
98
  print(embeddings.shape)
 
149
  - `per_device_train_batch_size`: 128
150
  - `per_device_eval_batch_size`: 128
151
  - `learning_rate`: 2e-05
152
+ - `num_train_epochs`: 15
153
  - `warmup_ratio`: 0.1
154
  - `fp16`: True
155
 
 
173
  - `adam_beta2`: 0.999
174
  - `adam_epsilon`: 1e-08
175
  - `max_grad_norm`: 1.0
176
+ - `num_train_epochs`: 15
177
  - `max_steps`: -1
178
  - `lr_scheduler_type`: linear
179
  - `lr_scheduler_kwargs`: {}
 
274
  ### Training Logs
275
  | Epoch | Step | Training Loss |
276
  |:------:|:----:|:-------------:|
277
+ | 0.1721 | 100 | 10.8697 |
278
+ | 0.3442 | 200 | 9.1125 |
279
+ | 0.5164 | 300 | 6.8873 |
280
+ | 0.6885 | 400 | 3.1124 |
281
+ | 0.8606 | 500 | 1.0882 |
282
+ | 1.0327 | 600 | 0.869 |
283
+ | 1.2048 | 700 | 0.6952 |
284
+ | 1.3769 | 800 | 0.5522 |
285
+ | 1.5491 | 900 | 0.5184 |
286
+ | 1.7212 | 1000 | 0.3996 |
287
 
288
 
289
  ### Framework Versions
checkpoint-1000/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea4b264b62f1f3dfc8756de31cee45d4ac202c2713bed99f962a11e9087a35e3
3
  size 90864192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73e5bceeeb8bd41387e7ebe4b365333a81a52ae62160e1fa7e0a03dfc37ba0e3
3
  size 90864192
checkpoint-1000/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a233a843b3a76d2b79bcb84775a6f946f363ec4c1cb82d0a2a2fb5336725f64
3
  size 180607738
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb34b407b9c534b08406d0b763420188b77a8c5ae59b33faef820746c1533433
3
  size 180607738
checkpoint-1000/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9b4e3b023cf818a92a7faebf3bfd08e4bf4825bcbd3669f92d1f25d60a14eef
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b320d6c3b1cd02590a033bf1d85ee904f31b839ca127ff50b770410fdc09a9e7
3
  size 14244
checkpoint-1000/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75a7920109b4e336482779ea4d3a308920045e01ee2548b630deb1a47f9a0f7a
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15dfa03098fd205b4cc1a5079e6c0db8a1fc9977130aed17d77d33eb6d658add
3
  size 1064
checkpoint-1000/trainer_state.json CHANGED
@@ -10,79 +10,79 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.1721170395869191,
13
- "grad_norm": 85.81483459472656,
14
- "learning_rate": 1.6581196581196585e-05,
15
- "loss": 8.9697,
16
  "step": 100
17
  },
18
  {
19
  "epoch": 0.3442340791738382,
20
- "grad_norm": 12.870137214660645,
21
- "learning_rate": 1.8468899521531104e-05,
22
- "loss": 2.6119,
23
  "step": 200
24
  },
25
  {
26
  "epoch": 0.5163511187607573,
27
- "grad_norm": 6.086756706237793,
28
- "learning_rate": 1.657416267942584e-05,
29
- "loss": 0.9351,
30
  "step": 300
31
  },
32
  {
33
  "epoch": 0.6884681583476764,
34
- "grad_norm": 4.553544998168945,
35
- "learning_rate": 1.4660287081339715e-05,
36
- "loss": 0.7049,
37
  "step": 400
38
  },
39
  {
40
  "epoch": 0.8605851979345955,
41
- "grad_norm": 150.66317749023438,
42
- "learning_rate": 1.274641148325359e-05,
43
- "loss": 0.394,
44
  "step": 500
45
  },
46
  {
47
  "epoch": 1.0327022375215147,
48
- "grad_norm": 2.234631061553955,
49
- "learning_rate": 1.0832535885167466e-05,
50
- "loss": 0.6792,
51
  "step": 600
52
  },
53
  {
54
  "epoch": 1.2048192771084336,
55
- "grad_norm": 3.470113515853882,
56
- "learning_rate": 8.91866028708134e-06,
57
- "loss": 0.715,
58
  "step": 700
59
  },
60
  {
61
  "epoch": 1.3769363166953528,
62
- "grad_norm": 2.741096019744873,
63
- "learning_rate": 7.004784688995215e-06,
64
- "loss": 0.5655,
65
  "step": 800
66
  },
67
  {
68
  "epoch": 1.549053356282272,
69
- "grad_norm": 1.5165340900421143,
70
- "learning_rate": 5.090909090909091e-06,
71
- "loss": 0.5354,
72
  "step": 900
73
  },
74
  {
75
  "epoch": 1.721170395869191,
76
- "grad_norm": 1.2075858116149902,
77
- "learning_rate": 3.177033492822967e-06,
78
- "loss": 0.6299,
79
  "step": 1000
80
  }
81
  ],
82
  "logging_steps": 100,
83
- "max_steps": 1162,
84
  "num_input_tokens_seen": 0,
85
- "num_train_epochs": 2,
86
  "save_steps": 500,
87
  "stateful_callbacks": {
88
  "TrainerControl": {
 
10
  "log_history": [
11
  {
12
  "epoch": 0.1721170395869191,
13
+ "grad_norm": 134.16664123535156,
14
+ "learning_rate": 2.224770642201835e-06,
15
+ "loss": 10.8697,
16
  "step": 100
17
  },
18
  {
19
  "epoch": 0.3442340791738382,
20
+ "grad_norm": 69.6041259765625,
21
+ "learning_rate": 4.5183486238532115e-06,
22
+ "loss": 9.1125,
23
  "step": 200
24
  },
25
  {
26
  "epoch": 0.5163511187607573,
27
+ "grad_norm": 103.4394760131836,
28
+ "learning_rate": 6.8119266055045875e-06,
29
+ "loss": 6.8873,
30
  "step": 300
31
  },
32
  {
33
  "epoch": 0.6884681583476764,
34
+ "grad_norm": 32.237342834472656,
35
+ "learning_rate": 9.08256880733945e-06,
36
+ "loss": 3.1124,
37
  "step": 400
38
  },
39
  {
40
  "epoch": 0.8605851979345955,
41
+ "grad_norm": 13.523033142089844,
42
+ "learning_rate": 1.1376146788990828e-05,
43
+ "loss": 1.0882,
44
  "step": 500
45
  },
46
  {
47
  "epoch": 1.0327022375215147,
48
+ "grad_norm": 56.13002395629883,
49
+ "learning_rate": 1.3669724770642203e-05,
50
+ "loss": 0.869,
51
  "step": 600
52
  },
53
  {
54
  "epoch": 1.2048192771084336,
55
+ "grad_norm": 4.41276741027832,
56
+ "learning_rate": 1.5963302752293578e-05,
57
+ "loss": 0.6952,
58
  "step": 700
59
  },
60
  {
61
  "epoch": 1.3769363166953528,
62
+ "grad_norm": 1.8771318197250366,
63
+ "learning_rate": 1.8256880733944955e-05,
64
+ "loss": 0.5522,
65
  "step": 800
66
  },
67
  {
68
  "epoch": 1.549053356282272,
69
+ "grad_norm": 5.131401062011719,
70
+ "learning_rate": 1.9938798928981258e-05,
71
+ "loss": 0.5184,
72
  "step": 900
73
  },
74
  {
75
  "epoch": 1.721170395869191,
76
+ "grad_norm": 3.2520999908447266,
77
+ "learning_rate": 1.9683794466403164e-05,
78
+ "loss": 0.3996,
79
  "step": 1000
80
  }
81
  ],
82
  "logging_steps": 100,
83
+ "max_steps": 8715,
84
  "num_input_tokens_seen": 0,
85
+ "num_train_epochs": 15,
86
  "save_steps": 500,
87
  "stateful_callbacks": {
88
  "TrainerControl": {
checkpoint-1000/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:57644da3fa51947d8e4d49382ff60305072c84a28b92b508a49ba8ad874ddd0f
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac8296029d0c3febd6057eb740de9cfe0b929fa172fd988f31151dfac15c657c
3
  size 5496