mkmkmkmk commited on
Commit
c8fb5cc
1 Parent(s): de6a880

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -62,7 +62,7 @@ We tuned learning rate and training epochs for each model and task following [th
62
  For tasks other than MARC-ja, the maximum length is short, so the attention_type was set to "original_full" and fine-tuning was performed for tasks other than MARC-ja. For MARC-ja, both "block_sparse" and "original_full" were used.
63
 
64
  | Model |MARC-ja(block_sparse)/acc | MARC-ja(original_full)/acc | JSTS/pearson | JSTS/spearman | JNLI/acc | JSQuAD/EM | JSQuAD/F1 | JComQA/acc |
65
- |-------------------------------|-------------|--------------|---------------|----------|-----------|-----------|------------|
66
  | Waseda RoBERTa base | - | 0.965 | 0.913 | 0.876 | 0.905 | 0.853 | 0.916 | 0.853 |
67
  | Waseda RoBERTa large (seq512) | - |0.969 | 0.925 | 0.890 | 0.928 | 0.910 | 0.955 | 0.900 |
68
  | BigBird base | 0.959 |0.959 | 0.888 | 0.846 | 0.896 | 0.884 | 0.933 | 0.787 |
 
62
  For tasks other than MARC-ja, the maximum length is short, so the attention_type was set to "original_full" and fine-tuning was performed for tasks other than MARC-ja. For MARC-ja, both "block_sparse" and "original_full" were used.
63
 
64
  | Model |MARC-ja(block_sparse)/acc | MARC-ja(original_full)/acc | JSTS/pearson | JSTS/spearman | JNLI/acc | JSQuAD/EM | JSQuAD/F1 | JComQA/acc |
65
+ |-------------------------------|-------------|--------------|---------------|----------|-----------|-----------|------------|------------|
66
  | Waseda RoBERTa base | - | 0.965 | 0.913 | 0.876 | 0.905 | 0.853 | 0.916 | 0.853 |
67
  | Waseda RoBERTa large (seq512) | - |0.969 | 0.925 | 0.890 | 0.928 | 0.910 | 0.955 | 0.900 |
68
  | BigBird base | 0.959 |0.959 | 0.888 | 0.846 | 0.896 | 0.884 | 0.933 | 0.787 |