|
--- |
|
license: mit |
|
task_categories: |
|
- zero-shot-classification |
|
size_categories: |
|
- n<1K |
|
--- |
|
|
|
# MMVP-VLM Benchmark Datacard |
|
|
|
## Basic Information |
|
|
|
**Title:** MMVP-VLM Benchmark |
|
|
|
**Description:** The MMVP-VLM (Multimodal Visual Patterns - Visual Language Models) Benchmark is designed to systematically evaluate the performance of recent CLIP-based models in understanding and processing visual patterns. It distills a subset of questions from the original MMVP benchmark into simpler language descriptions, categorizing them into distinct visual patterns. Each visual pattern is represented by 15 text-image pairs. The benchmark assesses whether CLIP models can accurately match these image-text combinations, providing insights into the capabilities and limitations of these models. |
|
|
|
## Dataset Details |
|
|
|
- **Content Types:** Text-Image Pairs |
|
- **Volume:** Balanced number of questions for each visual pattern, with each pattern represented by 15 pairs. |
|
- **Source of Data:** Subset from MMVP benchmark, supplemented with additional questions for balance |
|
- **Data Collection Method:** Distillation and categorization of questions from MMVP benchmark into simpler language |
|
|
|
## Usage |
|
|
|
### Intended Use |
|
|
|
- Evaluation of CLIP models' ability to understand and process various visual patterns. |
|
|