OpenCLIP
PyTorch
clip
DFN2B-CLIP-ViT-L-14 / open_clip_config.json
apf1's picture
Upload models
fa63e19
raw
history blame contribute delete
634 Bytes
{
"model_cfg": {
"embed_dim": 768,
"quick_gelu": true,
"vision_cfg": {
"image_size": 224,
"layers": 24,
"width": 1024,
"patch_size": 14
},
"text_cfg": {
"context_length": 77,
"vocab_size": 49408,
"width": 768,
"heads": 12,
"layers": 12
}
},
"preprocess_cfg": {
"mean": [
0.48145466,
0.4578275,
0.40821073
],
"std": [
0.26862954,
0.26130258,
0.27577711
]
}
}