A Dataset for Testing the Adversarial Robustness of AI-Generated Image Detectors
AI & ML interests
None defined yet.
Recent Activity
View all activity
Models and data for ReT: Recurrence-Enhanced Vision-and-Language Transformers for Robust Multimodal Document Retrieval [CVPR 2025]
LLaVA-MORE: A Comparative Study of LLMs and Visual Backbones
for Enhanced Visual Instruction Tuning
-
aimagelab/LLaVA_MORE-llama_3_1-8B-pretrain
Image-Text-to-Text • Updated • 5 -
aimagelab/LLaVA_MORE-llama_3_1-8B-finetuning
Image-Text-to-Text • 8B • Updated • 1.28k • 11 -
aimagelab/LLaVA_MORE-llama_3_1-8B-siglip-pretrain
Image-Text-to-Text • Updated • 14 -
aimagelab/LLaVA_MORE-llama_3_1-8B-siglip-finetuning
Image-Text-to-Text • 8B • Updated • 10 • 1
A Dataset for Testing the Adversarial Robustness of AI-Generated Image Detectors
Models and data for ReflectiVA: Augmenting Multimodal LLMs with Self-Reflective Tokens for Knowledge-based Visual Question Answering [CVPR 2025]
Models and data for ReT: Recurrence-Enhanced Vision-and-Language Transformers for Robust Multimodal Document Retrieval [CVPR 2025]
Models and dataset of Safe-CLIP: Removing NSFW Concepts from Vision-and-Language Models (https://arxiv.org/abs/2311.16254) [ECCV 2024]
LLaVA-MORE: A Comparative Study of LLMs and Visual Backbones
for Enhanced Visual Instruction Tuning
-
aimagelab/LLaVA_MORE-llama_3_1-8B-pretrain
Image-Text-to-Text • Updated • 5 -
aimagelab/LLaVA_MORE-llama_3_1-8B-finetuning
Image-Text-to-Text • 8B • Updated • 1.28k • 11 -
aimagelab/LLaVA_MORE-llama_3_1-8B-siglip-pretrain
Image-Text-to-Text • Updated • 14 -
aimagelab/LLaVA_MORE-llama_3_1-8B-siglip-finetuning
Image-Text-to-Text • 8B • Updated • 10 • 1