Upload folder using huggingface_hub
Browse files
checkpoint-1000/README.md
CHANGED
@@ -13,31 +13,31 @@ tags:
|
|
13 |
- dataset_size:74247
|
14 |
- loss:AnglELoss
|
15 |
widget:
|
16 |
-
- source_sentence:
|
17 |
-
sentences:
|
18 |
-
- matte concealer
|
19 |
-
- car refresher spray
|
20 |
-
- bread
|
21 |
-
- source_sentence: intensive antiwrinkle face cream
|
22 |
sentences:
|
23 |
- bread
|
24 |
-
- bread
|
25 |
- car refresher spray
|
26 |
-
-
|
|
|
27 |
sentences:
|
28 |
-
-
|
29 |
-
-
|
30 |
-
-
|
31 |
-
- source_sentence:
|
32 |
sentences:
|
33 |
-
-
|
34 |
-
-
|
|
|
|
|
|
|
|
|
35 |
- home scent
|
36 |
-
-
|
|
|
37 |
sentences:
|
38 |
-
-
|
39 |
-
-
|
40 |
-
-
|
41 |
---
|
42 |
|
43 |
# all-MiniLM-L6-v6-pair_score
|
@@ -90,9 +90,9 @@ from sentence_transformers import SentenceTransformer
|
|
90 |
model = SentenceTransformer("sentence_transformers_model_id")
|
91 |
# Run inference
|
92 |
sentences = [
|
93 |
-
'
|
94 |
-
'
|
95 |
-
'
|
96 |
]
|
97 |
embeddings = model.encode(sentences)
|
98 |
print(embeddings.shape)
|
@@ -149,7 +149,7 @@ You can finetune this model on your own dataset.
|
|
149 |
- `per_device_train_batch_size`: 128
|
150 |
- `per_device_eval_batch_size`: 128
|
151 |
- `learning_rate`: 2e-05
|
152 |
-
- `num_train_epochs`:
|
153 |
- `warmup_ratio`: 0.1
|
154 |
- `fp16`: True
|
155 |
|
@@ -173,7 +173,7 @@ You can finetune this model on your own dataset.
|
|
173 |
- `adam_beta2`: 0.999
|
174 |
- `adam_epsilon`: 1e-08
|
175 |
- `max_grad_norm`: 1.0
|
176 |
-
- `num_train_epochs`:
|
177 |
- `max_steps`: -1
|
178 |
- `lr_scheduler_type`: linear
|
179 |
- `lr_scheduler_kwargs`: {}
|
@@ -274,16 +274,16 @@ You can finetune this model on your own dataset.
|
|
274 |
### Training Logs
|
275 |
| Epoch | Step | Training Loss |
|
276 |
|:------:|:----:|:-------------:|
|
277 |
-
| 0.1721 | 100 |
|
278 |
-
| 0.3442 | 200 |
|
279 |
-
| 0.5164 | 300 |
|
280 |
-
| 0.6885 | 400 |
|
281 |
-
| 0.8606 | 500 |
|
282 |
-
| 1.0327 | 600 | 0.
|
283 |
-
| 1.2048 | 700 | 0.
|
284 |
-
| 1.3769 | 800 | 0.
|
285 |
-
| 1.5491 | 900 | 0.
|
286 |
-
| 1.7212 | 1000 | 0.
|
287 |
|
288 |
|
289 |
### Framework Versions
|
|
|
13 |
- dataset_size:74247
|
14 |
- loss:AnglELoss
|
15 |
widget:
|
16 |
+
- source_sentence: paste
|
|
|
|
|
|
|
|
|
|
|
17 |
sentences:
|
18 |
- bread
|
|
|
19 |
- car refresher spray
|
20 |
+
- hot drinks mug
|
21 |
+
- source_sentence: shoulder strap accessory for bags
|
22 |
sentences:
|
23 |
+
- home scent
|
24 |
+
- ball
|
25 |
+
- one outer pocket cross bag
|
26 |
+
- source_sentence: deep fried mozzarella sticks
|
27 |
sentences:
|
28 |
+
- one outer pocket cross bag
|
29 |
+
- insect killer
|
30 |
+
- stir fry vegetables bowl
|
31 |
+
- source_sentence: metal flower set tableaus
|
32 |
+
sentences:
|
33 |
+
- home scent
|
34 |
- home scent
|
35 |
+
- hygiene
|
36 |
+
- source_sentence: basic choker
|
37 |
sentences:
|
38 |
+
- hygiene
|
39 |
+
- unisex sweatshirt
|
40 |
+
- unisex sweatshirt
|
41 |
---
|
42 |
|
43 |
# all-MiniLM-L6-v6-pair_score
|
|
|
90 |
model = SentenceTransformer("sentence_transformers_model_id")
|
91 |
# Run inference
|
92 |
sentences = [
|
93 |
+
'basic choker',
|
94 |
+
'unisex sweatshirt',
|
95 |
+
'unisex sweatshirt',
|
96 |
]
|
97 |
embeddings = model.encode(sentences)
|
98 |
print(embeddings.shape)
|
|
|
149 |
- `per_device_train_batch_size`: 128
|
150 |
- `per_device_eval_batch_size`: 128
|
151 |
- `learning_rate`: 2e-05
|
152 |
+
- `num_train_epochs`: 15
|
153 |
- `warmup_ratio`: 0.1
|
154 |
- `fp16`: True
|
155 |
|
|
|
173 |
- `adam_beta2`: 0.999
|
174 |
- `adam_epsilon`: 1e-08
|
175 |
- `max_grad_norm`: 1.0
|
176 |
+
- `num_train_epochs`: 15
|
177 |
- `max_steps`: -1
|
178 |
- `lr_scheduler_type`: linear
|
179 |
- `lr_scheduler_kwargs`: {}
|
|
|
274 |
### Training Logs
|
275 |
| Epoch | Step | Training Loss |
|
276 |
|:------:|:----:|:-------------:|
|
277 |
+
| 0.1721 | 100 | 10.8697 |
|
278 |
+
| 0.3442 | 200 | 9.1125 |
|
279 |
+
| 0.5164 | 300 | 6.8873 |
|
280 |
+
| 0.6885 | 400 | 3.1124 |
|
281 |
+
| 0.8606 | 500 | 1.0882 |
|
282 |
+
| 1.0327 | 600 | 0.869 |
|
283 |
+
| 1.2048 | 700 | 0.6952 |
|
284 |
+
| 1.3769 | 800 | 0.5522 |
|
285 |
+
| 1.5491 | 900 | 0.5184 |
|
286 |
+
| 1.7212 | 1000 | 0.3996 |
|
287 |
|
288 |
|
289 |
### Framework Versions
|
checkpoint-1000/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 90864192
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:73e5bceeeb8bd41387e7ebe4b365333a81a52ae62160e1fa7e0a03dfc37ba0e3
|
3 |
size 90864192
|
checkpoint-1000/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 180607738
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cb34b407b9c534b08406d0b763420188b77a8c5ae59b33faef820746c1533433
|
3 |
size 180607738
|
checkpoint-1000/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b320d6c3b1cd02590a033bf1d85ee904f31b839ca127ff50b770410fdc09a9e7
|
3 |
size 14244
|
checkpoint-1000/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:15dfa03098fd205b4cc1a5079e6c0db8a1fc9977130aed17d77d33eb6d658add
|
3 |
size 1064
|
checkpoint-1000/trainer_state.json
CHANGED
@@ -10,79 +10,79 @@
|
|
10 |
"log_history": [
|
11 |
{
|
12 |
"epoch": 0.1721170395869191,
|
13 |
-
"grad_norm":
|
14 |
-
"learning_rate":
|
15 |
-
"loss":
|
16 |
"step": 100
|
17 |
},
|
18 |
{
|
19 |
"epoch": 0.3442340791738382,
|
20 |
-
"grad_norm":
|
21 |
-
"learning_rate":
|
22 |
-
"loss":
|
23 |
"step": 200
|
24 |
},
|
25 |
{
|
26 |
"epoch": 0.5163511187607573,
|
27 |
-
"grad_norm":
|
28 |
-
"learning_rate":
|
29 |
-
"loss":
|
30 |
"step": 300
|
31 |
},
|
32 |
{
|
33 |
"epoch": 0.6884681583476764,
|
34 |
-
"grad_norm":
|
35 |
-
"learning_rate":
|
36 |
-
"loss":
|
37 |
"step": 400
|
38 |
},
|
39 |
{
|
40 |
"epoch": 0.8605851979345955,
|
41 |
-
"grad_norm":
|
42 |
-
"learning_rate": 1.
|
43 |
-
"loss":
|
44 |
"step": 500
|
45 |
},
|
46 |
{
|
47 |
"epoch": 1.0327022375215147,
|
48 |
-
"grad_norm":
|
49 |
-
"learning_rate": 1.
|
50 |
-
"loss": 0.
|
51 |
"step": 600
|
52 |
},
|
53 |
{
|
54 |
"epoch": 1.2048192771084336,
|
55 |
-
"grad_norm":
|
56 |
-
"learning_rate":
|
57 |
-
"loss": 0.
|
58 |
"step": 700
|
59 |
},
|
60 |
{
|
61 |
"epoch": 1.3769363166953528,
|
62 |
-
"grad_norm":
|
63 |
-
"learning_rate":
|
64 |
-
"loss": 0.
|
65 |
"step": 800
|
66 |
},
|
67 |
{
|
68 |
"epoch": 1.549053356282272,
|
69 |
-
"grad_norm":
|
70 |
-
"learning_rate":
|
71 |
-
"loss": 0.
|
72 |
"step": 900
|
73 |
},
|
74 |
{
|
75 |
"epoch": 1.721170395869191,
|
76 |
-
"grad_norm":
|
77 |
-
"learning_rate":
|
78 |
-
"loss": 0.
|
79 |
"step": 1000
|
80 |
}
|
81 |
],
|
82 |
"logging_steps": 100,
|
83 |
-
"max_steps":
|
84 |
"num_input_tokens_seen": 0,
|
85 |
-
"num_train_epochs":
|
86 |
"save_steps": 500,
|
87 |
"stateful_callbacks": {
|
88 |
"TrainerControl": {
|
|
|
10 |
"log_history": [
|
11 |
{
|
12 |
"epoch": 0.1721170395869191,
|
13 |
+
"grad_norm": 134.16664123535156,
|
14 |
+
"learning_rate": 2.224770642201835e-06,
|
15 |
+
"loss": 10.8697,
|
16 |
"step": 100
|
17 |
},
|
18 |
{
|
19 |
"epoch": 0.3442340791738382,
|
20 |
+
"grad_norm": 69.6041259765625,
|
21 |
+
"learning_rate": 4.5183486238532115e-06,
|
22 |
+
"loss": 9.1125,
|
23 |
"step": 200
|
24 |
},
|
25 |
{
|
26 |
"epoch": 0.5163511187607573,
|
27 |
+
"grad_norm": 103.4394760131836,
|
28 |
+
"learning_rate": 6.8119266055045875e-06,
|
29 |
+
"loss": 6.8873,
|
30 |
"step": 300
|
31 |
},
|
32 |
{
|
33 |
"epoch": 0.6884681583476764,
|
34 |
+
"grad_norm": 32.237342834472656,
|
35 |
+
"learning_rate": 9.08256880733945e-06,
|
36 |
+
"loss": 3.1124,
|
37 |
"step": 400
|
38 |
},
|
39 |
{
|
40 |
"epoch": 0.8605851979345955,
|
41 |
+
"grad_norm": 13.523033142089844,
|
42 |
+
"learning_rate": 1.1376146788990828e-05,
|
43 |
+
"loss": 1.0882,
|
44 |
"step": 500
|
45 |
},
|
46 |
{
|
47 |
"epoch": 1.0327022375215147,
|
48 |
+
"grad_norm": 56.13002395629883,
|
49 |
+
"learning_rate": 1.3669724770642203e-05,
|
50 |
+
"loss": 0.869,
|
51 |
"step": 600
|
52 |
},
|
53 |
{
|
54 |
"epoch": 1.2048192771084336,
|
55 |
+
"grad_norm": 4.41276741027832,
|
56 |
+
"learning_rate": 1.5963302752293578e-05,
|
57 |
+
"loss": 0.6952,
|
58 |
"step": 700
|
59 |
},
|
60 |
{
|
61 |
"epoch": 1.3769363166953528,
|
62 |
+
"grad_norm": 1.8771318197250366,
|
63 |
+
"learning_rate": 1.8256880733944955e-05,
|
64 |
+
"loss": 0.5522,
|
65 |
"step": 800
|
66 |
},
|
67 |
{
|
68 |
"epoch": 1.549053356282272,
|
69 |
+
"grad_norm": 5.131401062011719,
|
70 |
+
"learning_rate": 1.9938798928981258e-05,
|
71 |
+
"loss": 0.5184,
|
72 |
"step": 900
|
73 |
},
|
74 |
{
|
75 |
"epoch": 1.721170395869191,
|
76 |
+
"grad_norm": 3.2520999908447266,
|
77 |
+
"learning_rate": 1.9683794466403164e-05,
|
78 |
+
"loss": 0.3996,
|
79 |
"step": 1000
|
80 |
}
|
81 |
],
|
82 |
"logging_steps": 100,
|
83 |
+
"max_steps": 8715,
|
84 |
"num_input_tokens_seen": 0,
|
85 |
+
"num_train_epochs": 15,
|
86 |
"save_steps": 500,
|
87 |
"stateful_callbacks": {
|
88 |
"TrainerControl": {
|
checkpoint-1000/training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5496
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ac8296029d0c3febd6057eb740de9cfe0b929fa172fd988f31151dfac15c657c
|
3 |
size 5496
|