chenghao xiao

gowitheflow

AI & ML interests

None yet

Recent Activity

updated a dataset about 8 hours ago
SeaLLMs/TrueFalse-Statements-multilingual
updated a dataset 9 days ago
SeaLLMs/FreshQA-multilingual-augmented
published a dataset 9 days ago
SeaLLMs/FreshQA-multilingual-augmented
View all activity

Organizations

Multimodal Art Projection's profile picture Massive Text Embedding Benchmark's profile picture SeaLLMs - Language Models for Southeast Asian Languages's profile picture Go with the Flow's profile picture AudioConFit's profile picture RAR-b's profile picture Pixel Linguist's profile picture LaymanRRG's profile picture ARC-2024's profile picture cipher-ling's profile picture rigour's profile picture BioLaySumm Shared Task at ACL's profile picture

gowitheflow's activity

upvoted an article about 1 month ago
view article
Article

MIEB: The Benchmark That Stress-Tests Image-Text Embeddings Like Never Before

By isaacchung and 2 others
14
reacted to merterbak's post with 🚀👀🔥 about 1 month ago
view post
Post
2132
Here’s a cool paper I found: “Massive Image Embedding Benchmark (MIEB).” It is a new tool to test how good image embedding models are. It has 130 different tasks grouped into 8 categories, like image search, classification, clustering similar images, answering questions based on images, and understanding documents. It even covers 38 different languages.

The authors tested 50 models and found that no single model was best at everything. Some models were great at recognizing text inside images but struggled to handle complicated tasks like matching images and text that appear together.

Paper: https://arxiv.org/pdf/2504.10471v1
Code: https://github.com/embeddings-benchmark/mteb
  • 2 replies
·