jorgeduardo13 commited on
Commit
0343a93
·
1 Parent(s): 713c775

Lo logramos de nuevo equipo Platzi! 🤗

Browse files
Files changed (4) hide show
  1. README.md +8 -7
  2. all_results.json +11 -11
  3. eval_results.json +7 -7
  4. train_results.json +5 -5
README.md CHANGED
@@ -2,6 +2,7 @@
2
  license: apache-2.0
3
  base_model: distilbert-base-uncased
4
  tags:
 
5
  - generated_from_trainer
6
  datasets:
7
  - glue
@@ -15,7 +16,7 @@ model-index:
15
  name: Text Classification
16
  type: text-classification
17
  dataset:
18
- name: glue
19
  type: glue
20
  config: mrpc
21
  split: validation
@@ -23,10 +24,10 @@ model-index:
23
  metrics:
24
  - name: Accuracy
25
  type: accuracy
26
- value: 0.6887254901960784
27
  - name: F1
28
  type: f1
29
- value: 0.777583187390543
30
  ---
31
 
32
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -34,11 +35,11 @@ should probably proofread and complete it, then remove this comment. -->
34
 
35
  # platzi-distilbert-model-similaritytexts-JorgeEnciso
36
 
37
- This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset.
38
  It achieves the following results on the evaluation set:
39
- - Loss: 0.6390
40
- - Accuracy: 0.6887
41
- - F1: 0.7776
42
 
43
  ## Model description
44
 
 
2
  license: apache-2.0
3
  base_model: distilbert-base-uncased
4
  tags:
5
+ - text-classification
6
  - generated_from_trainer
7
  datasets:
8
  - glue
 
16
  name: Text Classification
17
  type: text-classification
18
  dataset:
19
+ name: datasetX
20
  type: glue
21
  config: mrpc
22
  split: validation
 
24
  metrics:
25
  - name: Accuracy
26
  type: accuracy
27
+ value: 0.7083333333333334
28
  - name: F1
29
  type: f1
30
+ value: 0.8125984251968503
31
  ---
32
 
33
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
35
 
36
  # platzi-distilbert-model-similaritytexts-JorgeEnciso
37
 
38
+ This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the datasetX dataset.
39
  It achieves the following results on the evaluation set:
40
+ - Loss: 0.5871
41
+ - Accuracy: 0.7083
42
+ - F1: 0.8126
43
 
44
  ## Model description
45
 
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
- "epoch": 4.0,
3
- "eval_accuracy": 0.6838235294117647,
4
- "eval_f1": 0.8122270742358079,
5
- "eval_loss": 0.624229907989502,
6
- "eval_runtime": 2.4395,
7
- "eval_samples_per_second": 167.246,
8
- "eval_steps_per_second": 20.906,
9
- "train_loss": 0.6376693077336729,
10
- "train_runtime": 114.4375,
11
- "train_samples_per_second": 128.21,
12
- "train_steps_per_second": 16.044
13
  }
 
1
  {
2
+ "epoch": 5.0,
3
+ "eval_accuracy": 0.7083333333333334,
4
+ "eval_f1": 0.8125984251968503,
5
+ "eval_loss": 0.5871198773384094,
6
+ "eval_runtime": 2.5693,
7
+ "eval_samples_per_second": 158.801,
8
+ "eval_steps_per_second": 19.85,
9
+ "train_loss": 0.5974179269961023,
10
+ "train_runtime": 138.812,
11
+ "train_samples_per_second": 132.121,
12
+ "train_steps_per_second": 16.533
13
  }
eval_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 4.0,
3
- "eval_accuracy": 0.6838235294117647,
4
- "eval_f1": 0.8122270742358079,
5
- "eval_loss": 0.624229907989502,
6
- "eval_runtime": 2.4395,
7
- "eval_samples_per_second": 167.246,
8
- "eval_steps_per_second": 20.906
9
  }
 
1
  {
2
+ "epoch": 5.0,
3
+ "eval_accuracy": 0.7083333333333334,
4
+ "eval_f1": 0.8125984251968503,
5
+ "eval_loss": 0.5871198773384094,
6
+ "eval_runtime": 2.5693,
7
+ "eval_samples_per_second": 158.801,
8
+ "eval_steps_per_second": 19.85
9
  }
train_results.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "epoch": 4.0,
3
- "train_loss": 0.6376693077336729,
4
- "train_runtime": 114.4375,
5
- "train_samples_per_second": 128.21,
6
- "train_steps_per_second": 16.044
7
  }
 
1
  {
2
+ "epoch": 5.0,
3
+ "train_loss": 0.5974179269961023,
4
+ "train_runtime": 138.812,
5
+ "train_samples_per_second": 132.121,
6
+ "train_steps_per_second": 16.533
7
  }