--- library_name: transformers license: apache-2.0 datasets: - RussianNLP/rucola language: - ru base_model: - deepvk/RuModernBERT-small pipeline_tag: text-classification metrics: - accuracy - matthews_correlation model-index: - name: d0rj/RuModernBERT-small-rucola results: - task: type: text-classification dataset: name: RussianNLP/rucola type: RussianNLP/rucola metrics: - name: Acc type: accuracy value: 0.70 - name: MCC type: matthews_correlation value: 0.25 source: name: RuCoLA benchmark url: https://rucola-benchmark.com/leaderboard? --- # d0rj/RuModernBERT-small-rucola ## Usage Labels: "1" refers to "acceptable", while "0" corresponds to "unacceptable". ### Simple ```python from transformers import pipeline pipe = pipeline('text-classification', model='d0rj/RuModernBERT-small-rucola') pipe(["Мне предоставилась возможность все видеть, сам оставаясь незамеченным.", "Весной в лесу очень хорошо"]) >>> [{'label': 'LABEL_0', 'score': 0.5270525217056274}, >>> {'label': 'LABEL_1', 'score': 0.923351526260376}] ``` ### Using weights ```python import torch from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForSequenceClassification.from_pretrained("d0rj/RuModernBERT-small-rucola") tokenizer = AutoTokenizer.from_pretrained("d0rj/RuModernBERT-small-rucola") @torch.inference_mode() def predict(text: str | list[str], model = model, tokenizer = tokenizer) -> list[int]: inputs = tokenizer(text, return_tensors='pt', padding=True, truncation=True).to(model.device) outputs = model(**inputs) probs = torch.nn.functional.softmax(outputs.logits, dim=-1) return probs.cpu().argmax(dim=-1).numpy().tolist() predict(["Мне предоставилась возможность все видеть, сам оставаясь незамеченным.", "Весной в лесу очень хорошо"]) >>> [0, 1] ``` ## Metrics | name | accuracy | MCC | model size, params | | ---- | -------- | --- | ------------------ | | [d0rj/RuModernBERT-small-rucola](https://huggingface.co/d0rj/RuModernBERT-small-rucola) | 0.7 | 0.25 | 34.5M | | [RussianNLP/ruRoBERTa-large-rucola](https://huggingface.co/RussianNLP/ruRoBERTa-large-rucola) | 0.82 | 0.56 | 355M | ## Training See [raw Weights & Biases logs](https://wandb.ai/d0rj/rucola_small) or [simple report](https://wandb.ai/d0rj/rucola_small/reports/RuModernBERT-small-rucola--VmlldzoxMTcyNDgzNg).