badaoui HF Staff commited on
Commit
77612d1
·
verified ·
1 Parent(s): 61f05e8

Adding Neuron-optimized model files

Browse files

🤖 Neuron Export Bot: On behalf of [badaoui](https://huggingface.co/badaoui), adding AWS Neuron-optimized model files.
Neuron-optimized models can achieve high-performance inference on AWS Inferentia and Trainium chips. Learn more:
- [AWS Neuron Documentation](https://awsdocs-neuron.readthedocs-hosted.com)
- [🤗 Optimum Neuron Guide](https://huggingface.co/docs/optimum-neuron/index)

Files changed (4) hide show
  1. .gitattributes +1 -0
  2. README.md +3 -2
  3. config.json +30 -0
  4. model.neuron +3 -0
.gitattributes CHANGED
@@ -8,3 +8,4 @@
8
  *.onnx filter=lfs diff=lfs merge=lfs -text
9
  *.msgpack filter=lfs diff=lfs merge=lfs -text
10
  model.safetensors filter=lfs diff=lfs merge=lfs -text
 
 
8
  *.onnx filter=lfs diff=lfs merge=lfs -text
9
  *.msgpack filter=lfs diff=lfs merge=lfs -text
10
  model.safetensors filter=lfs diff=lfs merge=lfs -text
11
+ model.neuron filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -8,11 +8,12 @@ metrics:
8
  tags:
9
  - bert
10
  - language-model
11
- - flaubert
12
- - flue
13
  - french
14
  - flaubert-small
15
  - cased
 
16
  ---
17
 
18
  # FlauBERT: Unsupervised Language Model Pre-training for French
 
8
  tags:
9
  - bert
10
  - language-model
11
+ - flaubert
12
+ - flue
13
  - french
14
  - flaubert-small
15
  - cased
16
+ - neuron
17
  ---
18
 
19
  # FlauBERT: Unsupervised Language Model Pre-training for French
config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "amp": 1,
3
  "architectures": [
4
  "FlaubertWithLMHeadModel"
@@ -50,6 +51,32 @@
50
  "n_heads": 8,
51
  "n_langs": 1,
52
  "n_layers": 6,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
53
  "pad_index": 2,
54
  "pad_token_id": 2,
55
  "pre_norm": true,
@@ -63,6 +90,9 @@
63
  "summary_type": "first",
64
  "summary_use_proj": true,
65
  "tokens_per_batch": -1,
 
 
 
66
  "unk_index": 3,
67
  "use_lang_emb": true,
68
  "vocab_size": 68729,
 
1
  {
2
+ "_attn_implementation_autoset": true,
3
  "amp": 1,
4
  "architectures": [
5
  "FlaubertWithLMHeadModel"
 
51
  "n_heads": 8,
52
  "n_langs": 1,
53
  "n_layers": 6,
54
+ "neuron": {
55
+ "auto_cast": null,
56
+ "auto_cast_type": null,
57
+ "compiler_type": "neuronx-cc",
58
+ "compiler_version": "2.19.8089.0+8ab9f450",
59
+ "disable_fallback": false,
60
+ "disable_fast_relayout": false,
61
+ "dynamic_batch_size": false,
62
+ "inline_weights_to_neff": true,
63
+ "input_names": [
64
+ "input_ids",
65
+ "attention_mask",
66
+ "token_type_ids"
67
+ ],
68
+ "model_type": "flaubert",
69
+ "optlevel": "2",
70
+ "output_attentions": false,
71
+ "output_hidden_states": false,
72
+ "output_names": [
73
+ "logits"
74
+ ],
75
+ "static_batch_size": 1,
76
+ "static_sequence_length": 128,
77
+ "task": "fill-mask",
78
+ "tensor_parallel_size": 1
79
+ },
80
  "pad_index": 2,
81
  "pad_token_id": 2,
82
  "pre_norm": true,
 
90
  "summary_type": "first",
91
  "summary_use_proj": true,
92
  "tokens_per_batch": -1,
93
+ "torch_dtype": "float32",
94
+ "torchscript": true,
95
+ "transformers_version": "4.51.3",
96
  "unk_index": 3,
97
  "use_lang_emb": true,
98
  "vocab_size": 68729,
model.neuron ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15a48c135e21d866e4b13451d42b8c694e015ce6840fda268287f4b0a5587de8
3
+ size 333028477