Original Model Link : https://huggingface.co/apple/DFN5B-CLIP-ViT-H-14-378

name: DFN5B-CLIP-ViT-H-14-378-SAFETENSORS
base_model: laion/CLIP-ViT-H-14-laion2B-s32B-b79K
license: apple-amlr
pipeline_tag: zero-shot-image-classification
tags: 
- clip
- Apple
- OpenAI
size:
- 3946981876
- 3.95 GB
tasks:
 - contrastive image-text
 - vision
language: en
papers:
- https://arxiv.org/abs/2309.17425
datasets:
- CommonPool-12.8B
license_link: LICENS

DFN5B-CLIP-ViT-H-14-378-SAFETENSORS

CLIP ViT-H/14 trained on DFN-5B Data Filtering Network filtered from a 43B uncurated pool from CommonPool-12.8B + 30B public images.

Downloads last month
7
Safetensors
Model size
987M params
Tensor type
I64
·
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for exdysa/DFN5B-CLIP-ViT-H-14-378-SAFETENSORS

Finetuned
(8)
this model

Collection including exdysa/DFN5B-CLIP-ViT-H-14-378-SAFETENSORS