File size: 22,615 Bytes
db46fd3
ef0235e
db46fd3
 
 
 
 
 
 
b95aaaa
8f38d53
 
 
b95aaaa
8f38d53
 
 
b95aaaa
8f38d53
 
 
b95aaaa
8f38d53
b95aaaa
8f38d53
 
 
 
 
b95aaaa
8f38d53
b95aaaa
8f38d53
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b95aaaa
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bfbb173
db46fd3
 
b95aaaa
db46fd3
b95aaaa
 
30a4932
 
 
b95aaaa
 
b4d508d
 
 
 
 
 
 
 
 
 
 
 
 
b95aaaa
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
db46fd3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b95aaaa
db46fd3
 
 
 
b95aaaa
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8b680f4
 
30a4932
8b680f4
b95aaaa
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
---
license: cc-by-nc-sa-4.0
library_name: span-marker
tags:
- span-marker
- token-classification
- ner
- named-entity-recognition
pipeline_tag: token-classification
widget:
- text: "Amelia Earthart flog mit ihrer einmotorigen Lockheed Vega 5B über den Atlantik nach Paris."
  example_title: "German"
- text: "Amelia Earhart flew her single engine Lockheed Vega 5B across the Atlantic to Paris."
  example_title: "English"
- text: "Amelia Earthart voló su Lockheed Vega 5B monomotor a través del Océano Atlántico hasta París."
  example_title: "Spanish"
- text: "Amelia Earthart a fait voler son monomoteur Lockheed Vega 5B à travers l'ocean Atlantique jusqu'à Paris."
  example_title: "French"
- text: "Amelia Earhart ha volato con il suo monomotore Lockheed Vega 5B attraverso l'Atlantico fino a Parigi."
  example_title: "Italian"
- text: "Amelia Earthart vloog met haar één-motorige Lockheed Vega 5B over de Atlantische Oceaan naar Parijs."
  example_title: "Dutch"
- text: "Amelia Earthart przeleciała swoim jednosilnikowym samolotem Lockheed Vega 5B przez Ocean Atlantycki do Paryża."
  example_title: "Polish"
- text: "Amelia Earhart voou em seu monomotor Lockheed Vega 5B através do Atlântico para Paris."
  example_title: "Portuguese"
- text: "Амелия Эртхарт перелетела на своем одномоторном самолете Lockheed Vega 5B через Атлантический океан в Париж."
  example_title: "Russian"
- text: "Amelia Earthart flaug eins hreyfils Lockheed Vega 5B yfir Atlantshafið til Parísar."
  example_title: "Icelandic"
- text: "Η Amelia Earthart πέταξε το μονοκινητήριο Lockheed Vega 5B της πέρα ​​από τον Ατλαντικό Ωκεανό στο Παρίσι."
  example_title: "Greek"
- text: "Amelia Earhartová přeletěla se svým jednomotorovým Lockheed Vega 5B přes Atlantik do Paříže."
  example_title: "Czech"
- text: "Amelia Earhart lensi yksimoottorisella Lockheed Vega 5B:llä Atlantin yli Pariisiin."
  example_title: "Finnish"
- text: "Amelia Earhart fløj med sin enmotoriske Lockheed Vega 5B over Atlanten til Paris."
  example_title: "Danish"
- text: "Amelia Earhart flög sin enmotoriga Lockheed Vega 5B över Atlanten till Paris."
  example_title: "Swedish"
- text: "Amelia Earhart fløy sin enmotoriske Lockheed Vega 5B over Atlanterhavet til Paris."
  example_title: "Norwegian"
- text: "Amelia Earhart și-a zburat cu un singur motor Lockheed Vega 5B peste Atlantic până la Paris."
  example_title: "Romanian"
- text: "Amelia Earhart menerbangkan mesin tunggal Lockheed Vega 5B melintasi Atlantik ke Paris."
  example_title: "Indonesian"
- text: "Амелія Эрхарт пераляцела на сваім аднаматорным Lockheed Vega 5B праз Атлантыку ў Парыж."
  example_title: "Belarusian"
- text: "Амелія Ергарт перелетіла на своєму одномоторному літаку Lockheed Vega 5B через Атлантику до Парижа."
  example_title: "Ukrainian"
- text: "Amelia Earhart preletjela je svojim jednomotornim zrakoplovom Lockheed Vega 5B preko Atlantika do Pariza."
  example_title: "Croatian"
- text: "Amelia Earhart lendas oma ühemootoriga Lockheed Vega 5B üle Atlandi ookeani Pariisi ."
  example_title: "Estonian"
model-index:
  - name: SpanMarker w. bert-base-multilingual-cased on MultiNERD by Tom Aarsen
    results:
      - task:
          type: token-classification
          name: Named Entity Recognition
        dataset:
          type: Babelscape/multinerd
          name: MultiNERD
          split: test
          revision: 2814b78e7af4b5a1f1886fe7ad49632de4d9dd25
        metrics:
          - type: f1
            value: 0.92478
            name: F1
          - type: precision
            value: 0.93385
            name: Precision
          - type: recall
            value: 0.91588
            name: Recall
datasets:
  - Babelscape/multinerd
language:
  - multilingual
metrics:
  - f1
  - recall
  - precision
base_model: bert-base-multilingual-cased
---

# SpanMarker for Multilingual Named Entity Recognition

This is a [SpanMarker](https://github.com/tomaarsen/SpanMarkerNER) model that can be used for multilingual Named Entity Recognition trained on the [MultiNERD](https://huggingface.co/datasets/Babelscape/multinerd) dataset. In particular, this SpanMarker model uses [bert-base-multilingual-cased](https://huggingface.co/bert-base-multilingual-cased) as the underlying encoder. See [train.py](train.py) for the training script.

Is your data not (always) capitalized correctly? Then consider using this uncased variant of this model by [@lxyuan](https://huggingface.co/lxyuan) for better performance: 
[lxyuan/span-marker-bert-base-multilingual-uncased-multinerd](https://huggingface.co/lxyuan/span-marker-bert-base-multilingual-uncased-multinerd).

## Metrics

| **Language** | **Precision** | **Recall** | **F1**     |
|--------------|---------------|------------|------------|
| **all**      | 93.39         | 91.59      | **92.48**  |
| **de**       | 95.21         | 94.32      | **94.76**  |
| **en**       | 95.07         | 95.29      | **95.18**  |
| **es**       | 93.50         | 89.65      | **91.53**  |
| **fr**       | 93.86         | 90.07      | **91.92**  |
| **it**       | 91.63         | 93.57      | **92.59**  |
| **nl**       | 94.86         | 91.74      | **93.27**  |
| **pl**       | 93.51         | 91.83      | **92.66**  |
| **pt**       | 94.48         | 91.30      | **92.86**  |
| **ru**       | 93.70         | 93.10      | **93.39**  |
| **zh**       | 88.36         | 85.71      | **87.02**  |

## Label set

| Class | Description | Examples |
|-------|-------------|----------|
PER (person) | People | Ray Charles, Jessica Alba, Leonardo DiCaprio, Roger Federer, Anna Massey. |
ORG (organization) | Associations, companies, agencies, institutions, nationalities and religious or political groups | University of Edinburgh, San Francisco Giants, Google, Democratic Party. |
LOC (location) | Physical locations (e.g. mountains, bodies of water), geopolitical entities (e.g. cities, states), and facilities (e.g. bridges, buildings, airports). | Rome, Lake Paiku, Chrysler Building, Mount Rushmore, Mississippi River. |
ANIM (animal) | Breeds of dogs, cats and other animals, including their scientific names. | Maine Coon, African Wild Dog, Great White Shark, New Zealand Bellbird. |
BIO (biological) | Genus of fungus, bacteria and protoctists, families of viruses, and other biological entities. | Herpes Simplex Virus, Escherichia Coli, Salmonella, Bacillus Anthracis. |
CEL (celestial) | Planets, stars, asteroids, comets, nebulae, galaxies and other astronomical objects. | Sun, Neptune, Asteroid 187 Lamberta, Proxima Centauri, V838 Monocerotis. |
DIS (disease) | Physical, mental, infectious, non-infectious, deficiency, inherited, degenerative, social and self-inflicted diseases. | Alzheimer’s Disease, Cystic Fibrosis, Dilated Cardiomyopathy, Arthritis. |
EVE (event) | Sport events, battles, wars and other events. | American Civil War, 2003 Wimbledon Championships, Cannes Film Festival. |
FOOD (food) | Foods and drinks. | Carbonara, Sangiovese, Cheddar Beer Fondue, Pizza Margherita. |
INST (instrument) | Technological instruments, mechanical instruments, musical instruments, and other tools. | Spitzer Space Telescope, Commodore 64, Skype, Apple Watch, Fender Stratocaster. |
MEDIA (media) | Titles of films, books, magazines, songs and albums, fictional characters and languages. | Forbes, American Psycho, Kiss Me Once, Twin Peaks, Disney Adventures. |
PLANT (plant) | Types of trees, flowers, and other plants, including their scientific names. | Salix, Quercus Petraea, Douglas Fir, Forsythia, Artemisia Maritima. |
MYTH (mythological) | Mythological and religious entities. | Apollo, Persephone, Aphrodite, Saint Peter, Pope Gregory I, Hercules. |
TIME (time) | Specific and well-defined time intervals, such as eras, historical periods, centuries, years and important days. No months and days of the week. | Renaissance, Middle Ages, Christmas, Great Depression, 17th Century, 2012. |
VEHI (vehicle) | Cars, motorcycles and other vehicles. | Ferrari Testarossa, Suzuki Jimny, Honda CR-X, Boeing 747, Fairey Fulmar.

## Usage

To use this model for inference, first install the `span_marker` library:

```bash
pip install span_marker
```

You can then run inference with this model like so:

```python
from span_marker import SpanMarkerModel

# Download from the 🤗 Hub
model = SpanMarkerModel.from_pretrained("tomaarsen/span-marker-mbert-base-multinerd")
# Run inference
entities = model.predict("Amelia Earhart flew her single engine Lockheed Vega 5B across the Atlantic to Paris.")
```

See the [SpanMarker](https://github.com/tomaarsen/SpanMarkerNER) repository for documentation and additional information on this library.

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 1

### Training results

| Training Loss | Epoch | Step   | Validation Loss | Overall Precision | Overall Recall | Overall F1 | Overall Accuracy |
|:-------------:|:-----:|:------:|:---------------:|:-----------------:|:--------------:|:----------:|:----------------:|
| 0.0179        | 0.01  | 1000   | 0.0146          | 0.8101            | 0.7616         | 0.7851     | 0.9530           |
| 0.0099        | 0.02  | 2000   | 0.0091          | 0.8571            | 0.8425         | 0.8498     | 0.9663           |
| 0.0085        | 0.03  | 3000   | 0.0078          | 0.8729            | 0.8579         | 0.8653     | 0.9700           |
| 0.0075        | 0.04  | 4000   | 0.0072          | 0.8821            | 0.8724         | 0.8772     | 0.9739           |
| 0.0074        | 0.05  | 5000   | 0.0075          | 0.8622            | 0.8841         | 0.8730     | 0.9722           |
| 0.0074        | 0.06  | 6000   | 0.0067          | 0.9056            | 0.8568         | 0.8805     | 0.9749           |
| 0.0066        | 0.07  | 7000   | 0.0065          | 0.9082            | 0.8543         | 0.8804     | 0.9737           |
| 0.0063        | 0.08  | 8000   | 0.0066          | 0.9039            | 0.8617         | 0.8823     | 0.9745           |
| 0.0062        | 0.09  | 9000   | 0.0062          | 0.9323            | 0.8425         | 0.8852     | 0.9754           |
| 0.007         | 0.1   | 10000  | 0.0066          | 0.8898            | 0.8758         | 0.8827     | 0.9746           |
| 0.006         | 0.11  | 11000  | 0.0061          | 0.8986            | 0.8841         | 0.8913     | 0.9766           |
| 0.006         | 0.12  | 12000  | 0.0061          | 0.9171            | 0.8628         | 0.8891     | 0.9763           |
| 0.0062        | 0.13  | 13000  | 0.0060          | 0.9264            | 0.8634         | 0.8938     | 0.9772           |
| 0.0059        | 0.14  | 14000  | 0.0059          | 0.9323            | 0.8508         | 0.8897     | 0.9763           |
| 0.0059        | 0.15  | 15000  | 0.0060          | 0.9011            | 0.8815         | 0.8912     | 0.9758           |
| 0.0059        | 0.16  | 16000  | 0.0060          | 0.9221            | 0.8598         | 0.8898     | 0.9763           |
| 0.0056        | 0.17  | 17000  | 0.0058          | 0.9098            | 0.8839         | 0.8967     | 0.9775           |
| 0.0055        | 0.18  | 18000  | 0.0060          | 0.9103            | 0.8739         | 0.8917     | 0.9765           |
| 0.0054        | 0.19  | 19000  | 0.0056          | 0.9135            | 0.8726         | 0.8925     | 0.9774           |
| 0.0052        | 0.2   | 20000  | 0.0058          | 0.9108            | 0.8834         | 0.8969     | 0.9773           |
| 0.0053        | 0.21  | 21000  | 0.0058          | 0.9038            | 0.8866         | 0.8951     | 0.9773           |
| 0.0057        | 0.22  | 22000  | 0.0057          | 0.9130            | 0.8762         | 0.8942     | 0.9775           |
| 0.0056        | 0.23  | 23000  | 0.0053          | 0.9375            | 0.8604         | 0.8973     | 0.9781           |
| 0.005         | 0.24  | 24000  | 0.0054          | 0.9253            | 0.8822         | 0.9032     | 0.9784           |
| 0.0055        | 0.25  | 25000  | 0.0055          | 0.9182            | 0.8807         | 0.8991     | 0.9787           |
| 0.0049        | 0.26  | 26000  | 0.0053          | 0.9311            | 0.8702         | 0.8997     | 0.9783           |
| 0.0051        | 0.27  | 27000  | 0.0054          | 0.9192            | 0.8877         | 0.9032     | 0.9787           |
| 0.0051        | 0.28  | 28000  | 0.0053          | 0.9332            | 0.8783         | 0.9049     | 0.9795           |
| 0.0049        | 0.29  | 29000  | 0.0054          | 0.9311            | 0.8672         | 0.8981     | 0.9789           |
| 0.0047        | 0.3   | 30000  | 0.0054          | 0.9165            | 0.8954         | 0.9058     | 0.9796           |
| 0.005         | 0.31  | 31000  | 0.0052          | 0.9079            | 0.9016         | 0.9047     | 0.9787           |
| 0.0051        | 0.32  | 32000  | 0.0051          | 0.9157            | 0.9001         | 0.9078     | 0.9796           |
| 0.0046        | 0.33  | 33000  | 0.0051          | 0.9147            | 0.8935         | 0.9040     | 0.9788           |
| 0.0046        | 0.34  | 34000  | 0.0050          | 0.9229            | 0.8847         | 0.9034     | 0.9793           |
| 0.005         | 0.35  | 35000  | 0.0051          | 0.9198            | 0.8922         | 0.9058     | 0.9796           |
| 0.0047        | 0.36  | 36000  | 0.0050          | 0.9321            | 0.8890         | 0.9100     | 0.9807           |
| 0.0048        | 0.37  | 37000  | 0.0050          | 0.9046            | 0.9133         | 0.9089     | 0.9800           |
| 0.0046        | 0.38  | 38000  | 0.0051          | 0.9170            | 0.8973         | 0.9071     | 0.9806           |
| 0.0048        | 0.39  | 39000  | 0.0050          | 0.9417            | 0.8775         | 0.9084     | 0.9805           |
| 0.0042        | 0.4   | 40000  | 0.0049          | 0.9238            | 0.8937         | 0.9085     | 0.9797           |
| 0.0038        | 0.41  | 41000  | 0.0048          | 0.9371            | 0.8920         | 0.9140     | 0.9812           |
| 0.0042        | 0.42  | 42000  | 0.0048          | 0.9359            | 0.8862         | 0.9104     | 0.9808           |
| 0.0051        | 0.43  | 43000  | 0.0049          | 0.9080            | 0.9060         | 0.9070     | 0.9805           |
| 0.0037        | 0.44  | 44000  | 0.0049          | 0.9328            | 0.8877         | 0.9097     | 0.9801           |
| 0.0041        | 0.45  | 45000  | 0.0049          | 0.9231            | 0.8975         | 0.9101     | 0.9813           |
| 0.0046        | 0.46  | 46000  | 0.0046          | 0.9308            | 0.8943         | 0.9122     | 0.9812           |
| 0.0038        | 0.47  | 47000  | 0.0047          | 0.9291            | 0.8969         | 0.9127     | 0.9815           |
| 0.0043        | 0.48  | 48000  | 0.0046          | 0.9308            | 0.8909         | 0.9104     | 0.9804           |
| 0.0043        | 0.49  | 49000  | 0.0046          | 0.9278            | 0.8954         | 0.9113     | 0.9800           |
| 0.0039        | 0.5   | 50000  | 0.0047          | 0.9173            | 0.9073         | 0.9123     | 0.9817           |
| 0.0043        | 0.51  | 51000  | 0.0045          | 0.9347            | 0.8962         | 0.9150     | 0.9821           |
| 0.0047        | 0.52  | 52000  | 0.0045          | 0.9266            | 0.9016         | 0.9139     | 0.9810           |
| 0.0035        | 0.53  | 53000  | 0.0046          | 0.9165            | 0.9122         | 0.9144     | 0.9820           |
| 0.0038        | 0.54  | 54000  | 0.0046          | 0.9231            | 0.9050         | 0.9139     | 0.9823           |
| 0.0036        | 0.55  | 55000  | 0.0046          | 0.9331            | 0.9005         | 0.9165     | 0.9828           |
| 0.0037        | 0.56  | 56000  | 0.0047          | 0.9246            | 0.9016         | 0.9129     | 0.9821           |
| 0.0035        | 0.57  | 57000  | 0.0044          | 0.9351            | 0.9003         | 0.9174     | 0.9829           |
| 0.0043        | 0.57  | 58000  | 0.0043          | 0.9257            | 0.9079         | 0.9167     | 0.9826           |
| 0.004         | 0.58  | 59000  | 0.0043          | 0.9286            | 0.9065         | 0.9174     | 0.9823           |
| 0.0041        | 0.59  | 60000  | 0.0044          | 0.9324            | 0.9050         | 0.9185     | 0.9825           |
| 0.0039        | 0.6   | 61000  | 0.0044          | 0.9268            | 0.9041         | 0.9153     | 0.9815           |
| 0.0038        | 0.61  | 62000  | 0.0043          | 0.9367            | 0.8918         | 0.9137     | 0.9819           |
| 0.0037        | 0.62  | 63000  | 0.0044          | 0.9249            | 0.9160         | 0.9205     | 0.9833           |
| 0.0036        | 0.63  | 64000  | 0.0043          | 0.9398            | 0.8975         | 0.9181     | 0.9827           |
| 0.0036        | 0.64  | 65000  | 0.0043          | 0.9260            | 0.9118         | 0.9188     | 0.9829           |
| 0.0035        | 0.65  | 66000  | 0.0044          | 0.9375            | 0.8988         | 0.9178     | 0.9828           |
| 0.0034        | 0.66  | 67000  | 0.0043          | 0.9272            | 0.9143         | 0.9207     | 0.9833           |
| 0.0033        | 0.67  | 68000  | 0.0044          | 0.9332            | 0.9024         | 0.9176     | 0.9827           |
| 0.0035        | 0.68  | 69000  | 0.0044          | 0.9396            | 0.8981         | 0.9184     | 0.9825           |
| 0.0038        | 0.69  | 70000  | 0.0042          | 0.9265            | 0.9163         | 0.9214     | 0.9827           |
| 0.0035        | 0.7   | 71000  | 0.0044          | 0.9375            | 0.9013         | 0.9191     | 0.9827           |
| 0.0037        | 0.71  | 72000  | 0.0042          | 0.9264            | 0.9171         | 0.9217     | 0.9830           |
| 0.0039        | 0.72  | 73000  | 0.0043          | 0.9399            | 0.9003         | 0.9197     | 0.9826           |
| 0.0039        | 0.73  | 74000  | 0.0041          | 0.9341            | 0.9094         | 0.9216     | 0.9832           |
| 0.0035        | 0.74  | 75000  | 0.0042          | 0.9301            | 0.9160         | 0.9230     | 0.9837           |
| 0.0037        | 0.75  | 76000  | 0.0042          | 0.9342            | 0.9107         | 0.9223     | 0.9835           |
| 0.0034        | 0.76  | 77000  | 0.0042          | 0.9331            | 0.9118         | 0.9223     | 0.9836           |
| 0.003         | 0.77  | 78000  | 0.0041          | 0.9330            | 0.9135         | 0.9231     | 0.9838           |
| 0.0034        | 0.78  | 79000  | 0.0041          | 0.9308            | 0.9082         | 0.9193     | 0.9832           |
| 0.0037        | 0.79  | 80000  | 0.0040          | 0.9346            | 0.9128         | 0.9236     | 0.9839           |
| 0.0032        | 0.8   | 81000  | 0.0041          | 0.9389            | 0.9128         | 0.9257     | 0.9841           |
| 0.0031        | 0.81  | 82000  | 0.0040          | 0.9293            | 0.9163         | 0.9227     | 0.9836           |
| 0.0032        | 0.82  | 83000  | 0.0041          | 0.9305            | 0.9160         | 0.9232     | 0.9835           |
| 0.0034        | 0.83  | 84000  | 0.0041          | 0.9327            | 0.9118         | 0.9221     | 0.9838           |
| 0.0028        | 0.84  | 85000  | 0.0041          | 0.9279            | 0.9216         | 0.9247     | 0.9839           |
| 0.0031        | 0.85  | 86000  | 0.0041          | 0.9326            | 0.9167         | 0.9246     | 0.9838           |
| 0.0029        | 0.86  | 87000  | 0.0040          | 0.9354            | 0.9158         | 0.9255     | 0.9841           |
| 0.0031        | 0.87  | 88000  | 0.0041          | 0.9327            | 0.9156         | 0.9241     | 0.9840           |
| 0.0033        | 0.88  | 89000  | 0.0040          | 0.9367            | 0.9141         | 0.9253     | 0.9846           |
| 0.0031        | 0.89  | 90000  | 0.0040          | 0.9379            | 0.9141         | 0.9259     | 0.9844           |
| 0.0031        | 0.9   | 91000  | 0.0040          | 0.9297            | 0.9184         | 0.9240     | 0.9843           |
| 0.0034        | 0.91  | 92000  | 0.0040          | 0.9299            | 0.9188         | 0.9243     | 0.9843           |
| 0.0036        | 0.92  | 93000  | 0.0039          | 0.9324            | 0.9175         | 0.9249     | 0.9843           |
| 0.0028        | 0.93  | 94000  | 0.0039          | 0.9399            | 0.9135         | 0.9265     | 0.9848           |
| 0.0029        | 0.94  | 95000  | 0.0040          | 0.9342            | 0.9173         | 0.9257     | 0.9845           |
| 0.003         | 0.95  | 96000  | 0.0040          | 0.9378            | 0.9184         | 0.9280     | 0.9850           |
| 0.0029        | 0.96  | 97000  | 0.0039          | 0.9380            | 0.9152         | 0.9264     | 0.9847           |
| 0.003         | 0.97  | 98000  | 0.0039          | 0.9372            | 0.9156         | 0.9263     | 0.9849           |
| 0.003         | 0.98  | 99000  | 0.0039          | 0.9387            | 0.9167         | 0.9276     | 0.9851           |
| 0.0031        | 0.99  | 100000 | 0.0039          | 0.9373            | 0.9177         | 0.9274     | 0.9849           |

### Framework versions

- SpanMarker 1.2.4
- Transformers 4.28.1
- Pytorch 1.13.1+cu117
- Datasets 2.12.0
- Tokenizers 0.13.2

## See also
* [lxyuan/span-marker-bert-base-multilingual-cased-multinerd](https://huggingface.co/lxyuan/span-marker-bert-base-multilingual-cased-multinerd) is similar to this model, but trained on 3 epochs instead of 2. It reaches better performance on 7 out of the 10 languages.
* [lxyuan/span-marker-bert-base-multilingual-uncased-multinerd](https://huggingface.co/lxyuan/span-marker-bert-base-multilingual-uncased-multinerd) is a strong uncased variant of this model, also trained on 3 epochs instead of 2.

## Contributions
Many thanks to [Simone Tedeschi](https://huggingface.co/sted97) from [Babelscape](https://babelscape.com) for his insight when training this model and his involvement in the creation of the training dataset.