JUHE API Marketplace
DATASET
Open Source Community

UniSim-Bench

UniSim‑Bench is a multimodal perception similarity benchmark created by New York University and EPFL, containing seven multimodal perception similarity tasks across 25 datasets. It covers various image‑to‑text tasks and is designed to evaluate model generalisation across tasks. The benchmark aggregates existing perception tasks and trains models using multi‑task learning. UniSim‑Bench is widely used to assess and improve multimodal perception models, especially for cross‑modal similarity evaluation and generative model quality assessment.

Updated 12/14/2024
arXiv

Description

Dataset Overview

Dataset Name

UniSim‑Bench

Dataset Description

UniSim‑Bench is a comprehensive benchmark covering 7 multimodal perception similarity tasks and 25 datasets. The benchmark is used to evaluate the performance of multimodal perception models and supports model training and evaluation.

Dataset Composition

  • Core 2AFC tasks: some datasets are used to train the UniSim model.
  • OOD generalisation tasks: all datasets are used only for testing.

Dataset Download

UniSim‑Bench can be found on HuggingFace at this link.

AI studio

Generate PPTs instantly with Nano Banana Pro.

Generate PPT Now

Access Dataset

Login to Access

Please login to view download links and access full dataset details.

Topics

Multimodal Perception
Model Evaluation

Source

Organization: arXiv

Created: 12/14/2024

Power Your Data Analysis with Premium AI Models

Supporting GPT-5, Claude-4, DeepSeek v3, Gemini and more.

Enjoy a free trial and save 20%+ compared to official pricing.