UniSim-Bench
UniSim‑Bench is a multimodal perception similarity benchmark created by New York University and EPFL, containing seven multimodal perception similarity tasks across 25 datasets. It covers various image‑to‑text tasks and is designed to evaluate model generalisation across tasks. The benchmark aggregates existing perception tasks and trains models using multi‑task learning. UniSim‑Bench is widely used to assess and improve multimodal perception models, especially for cross‑modal similarity evaluation and generative model quality assessment.
Description
Dataset Overview
Dataset Name
UniSim‑Bench
Dataset Description
UniSim‑Bench is a comprehensive benchmark covering 7 multimodal perception similarity tasks and 25 datasets. The benchmark is used to evaluate the performance of multimodal perception models and supports model training and evaluation.
Dataset Composition
- Core 2AFC tasks: some datasets are used to train the UniSim model.
- OOD generalisation tasks: all datasets are used only for testing.
Dataset Download
UniSim‑Bench can be found on HuggingFace at this link.
AI studio
Generate PPTs instantly with Nano Banana Pro.
Generate PPT NowAccess Dataset
Please login to view download links and access full dataset details.
Topics
Source
Organization: arXiv
Created: 12/14/2024
Power Your Data Analysis with Premium AI Models
Supporting GPT-5, Claude-4, DeepSeek v3, Gemini and more.
Enjoy a free trial and save 20%+ compared to official pricing.