clip-finetuned-csu-p14-336-e2l57-l

This model is a fine-tuned version of openai/clip-vit-large-patch14-336 on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 2.0096

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 128
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 2.0

Training results

Training Loss Epoch Step Validation Loss
0.7207 0.0921 500 2.0785
0.702 0.1842 1000 2.0786
0.7058 0.2763 1500 2.0656
0.7117 0.3685 2000 2.1134
0.7163 0.4606 2500 2.0307
0.6794 0.5527 3000 2.6134
0.6818 0.6448 3500 2.0862
0.6693 0.7369 4000 2.0295
0.6798 0.8290 4500 2.1399
0.6915 0.9211 5000 2.0116
0.6623 1.0133 5500 2.1612
0.6803 1.1054 6000 2.0493
0.6729 1.1975 6500 2.0518
0.6851 1.2896 7000 2.0391
0.6686 1.3817 7500 2.0697
0.6637 1.4738 8000 2.0457
0.6331 1.5660 8500 2.1388
0.6529 1.6581 9000 2.0182
0.6377 1.7502 9500 2.0096
0.6207 1.8423 10000 2.0490
0.6626 1.9344 10500 2.0149

Framework versions

  • Transformers 4.45.0.dev0
  • Pytorch 1.12.1
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
17
Safetensors
Model size
428M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for kevinoli/clip-finetuned-csu-p14-336-e2l57-l

Finetuned
(19)
this model