Mihaj commited on
Commit
6ea161b
·
verified ·
1 Parent(s): 0e82e65

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_name_or_path": "facebook/mms-1b-all",
3
- "activation_dropout": 0.05,
4
  "adapter_attn_dim": 16,
5
  "adapter_kernel_size": 3,
6
  "adapter_stride": 2,
@@ -9,7 +9,7 @@
9
  "architectures": [
10
  "Wav2Vec2ForCTC"
11
  ],
12
- "attention_dropout": 0.0,
13
  "bos_token_id": 1,
14
  "classifier_proj_size": 256,
15
  "codevector_dim": 1024,
@@ -50,22 +50,22 @@
50
  "feat_extract_activation": "gelu",
51
  "feat_extract_dropout": 0.0,
52
  "feat_extract_norm": "layer",
53
- "feat_proj_dropout": 0.0,
54
  "feat_quantizer_dropout": 0.0,
55
  "final_dropout": 0.05,
56
  "hidden_act": "gelu",
57
- "hidden_dropout": 0.0,
58
  "hidden_size": 1280,
59
  "initializer_range": 0.02,
60
  "intermediate_size": 5120,
61
  "layer_norm_eps": 1e-05,
62
- "layerdrop": 0.0,
63
  "mask_feature_length": 10,
64
  "mask_feature_min_masks": 0,
65
  "mask_feature_prob": 0.0,
66
  "mask_time_length": 10,
67
  "mask_time_min_masks": 2,
68
- "mask_time_prob": 0.05,
69
  "model_type": "wav2vec2",
70
  "num_adapter_layers": 3,
71
  "num_attention_heads": 16,
 
1
  {
2
  "_name_or_path": "facebook/mms-1b-all",
3
+ "activation_dropout": 0.055,
4
  "adapter_attn_dim": 16,
5
  "adapter_kernel_size": 3,
6
  "adapter_stride": 2,
 
9
  "architectures": [
10
  "Wav2Vec2ForCTC"
11
  ],
12
+ "attention_dropout": 0.094,
13
  "bos_token_id": 1,
14
  "classifier_proj_size": 256,
15
  "codevector_dim": 1024,
 
50
  "feat_extract_activation": "gelu",
51
  "feat_extract_dropout": 0.0,
52
  "feat_extract_norm": "layer",
53
+ "feat_proj_dropout": 0.04,
54
  "feat_quantizer_dropout": 0.0,
55
  "final_dropout": 0.05,
56
  "hidden_act": "gelu",
57
+ "hidden_dropout": 0.047,
58
  "hidden_size": 1280,
59
  "initializer_range": 0.02,
60
  "intermediate_size": 5120,
61
  "layer_norm_eps": 1e-05,
62
+ "layerdrop": 0.041,
63
  "mask_feature_length": 10,
64
  "mask_feature_min_masks": 0,
65
  "mask_feature_prob": 0.0,
66
  "mask_time_length": 10,
67
  "mask_time_min_masks": 2,
68
+ "mask_time_prob": 0.082,
69
  "model_type": "wav2vec2",
70
  "num_adapter_layers": 3,
71
  "num_attention_heads": 16,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c27fe8d4fdf56af9a29efecf6958f7e203ef2a77659700ddcbf470b294a17199
3
  size 1929521404
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78bf7e886ac5bde4f377cbc8c8dd7c2f9568f12834edc14b0b14b333740397ba
3
  size 1929521404
runs/May29_17-14-10_maxim-ubuntu3080/events.out.tfevents.1716992148.maxim-ubuntu3080.295161.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4073e7f2796690f96fb14f12646844fb2993f3bec4b6d234d87984f20ae32704
3
+ size 6812
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18706134b4881a89d17c04f3d1005ba72900c793a83648ef5cd9c88dba08d049
3
  size 6648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b93558dcd67db4933de255e66d06c0fac02381b35a986a5dc748cf2d3846918
3
  size 6648