Roberta-Base fine-tuned on PubMed Abstract
We limit the training textual data to the following MeSH
- All the child MeSH of
Biomarkers, Tumor(D014408)
, including things likeCarcinoembryonic Antigen(D002272)
- All the child MeSH of
Carcinoma(D002277)
, including things like all kinds of carcinoma: likeCarcinoma, Lewis Lung(D018827)
etc. around 80 kinds of carcinoma - All the child MeSH of
Clinical Trial(D016439)
- The training text file amounts to 531Mb
Training
- Trained on language modeling task, with
mlm_probability=0.15
, on 2 Tesla V100 32G
training_args = TrainingArguments(
output_dir=config.save, #select model path for checkpoint
overwrite_output_dir=True,
num_train_epochs=3,
per_device_train_batch_size=30,
per_device_eval_batch_size=60,
evaluation_strategy= 'steps',
save_total_limit=2,
eval_steps=250,
metric_for_best_model='eval_loss',
greater_is_better=False,
load_best_model_at_end =True,
prediction_loss_only=True,
report_to = "none")
- Downloads last month
- 152
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.