FastVLM: Efficient Vision Encoding for Vision Language Models Paper β’ 2412.13303 β’ Published Dec 17, 2024 β’ 21 β’ 4
FastVLM: Efficient Vision Encoding for Vision Language Models Paper β’ 2412.13303 β’ Published Dec 17, 2024 β’ 21
FastViT: A Fast Hybrid Vision Transformer using Structural Reparameterization Paper β’ 2303.14189 β’ Published Mar 24, 2023 β’ 4
MobileOne: An Improved One millisecond Mobile Backbone Paper β’ 2206.04040 β’ Published Jun 8, 2022
SAM-CLIP: Merging Vision Foundation Models towards Semantic and Spatial Understanding Paper β’ 2310.15308 β’ Published Oct 23, 2023 β’ 23
MobileCLIP: Fast Image-Text Models through Multi-Modal Reinforced Training Paper β’ 2311.17049 β’ Published Nov 28, 2023 β’ 2
Graph-Based Captioning: Enhancing Visual Descriptions by Interconnecting Region Captions Paper β’ 2407.06723 β’ Published Jul 9, 2024 β’ 11
Dataset Decomposition: Faster LLM Training with Variable Sequence Length Curriculum Paper β’ 2405.13226 β’ Published May 21, 2024 β’ 1
CLIP with Quality Captions: A Strong Pretraining for Vision Tasks Paper β’ 2405.08911 β’ Published May 14, 2024 β’ 1
view article Article WWDC 24: Running Mistral 7B with Core ML By FL33TW00D-HF and 3 others β’ Jul 22, 2024 β’ 61
MobileCLIP Models + DataCompDR Data Collection MobileCLIP: Mobile-friendly image-text models with SOTA zero-shot capabilities. DataCompDR: Improved datasets for training image-text SOTA models. β’ 22 items β’ Updated Oct 4, 2024 β’ 29