metadata
license: cc-by-4.0
task_categories:
- robotics
- visual-question-answering
language:
- en
size_categories:
- 1K<n<10K
configs:
- config_name: benchmark
data_files:
- split: single_arm
path: 3_generalized_planning/cross_embodiment/single_arm/questions.json
RoboBench: A Comprehensive Evaluation Benchmark for Multimodal Large Language Models as Embodied Brain
π Overview
RoboBench is a comprehensive evaluation benchmark designed to assess the capabilities of Multimodal Large Language Models (MLLMs) in embodied intelligence tasks. This benchmark provides a systematic framework for evaluating how well these models can understand and reason about robotic scenarios.
π― Key Features
- π§ Comprehensive Evaluation: Covers multiple aspects of embodied intelligence
- π Rich Dataset: Contains thousands of carefully curated examples
- π¬ Scientific Rigor: Designed with research-grade evaluation metrics
- π Multimodal: Supports text, images, and video data
- π€ Robotics Focus: Specifically tailored for robotic applications
π Dataset Statistics
| Category | Count | Description |
|---|---|---|
| Total Samples | 6092 | Comprehensive evaluation dataset |
| Image Samples | 1400 | High-quality visual data |
| Video Samples | 3142 | Temporal & Planning reasoning examples |
ποΈ Dataset Structure
RoboBench/
βββ 1_instruction_comprehension/ # Instruction understanding tasks
βββ 2_perception_reasoning/ # Visual perception and reasoning
βββ 3_generalized_planning/ # Cross-domain planning tasks
βββ 4_affordance_reasoning/ # Object affordance understanding
βββ 5_error_analysis/ # Error analysis and debugging
βββsystem_prompt.json. # Every task system prompts
π¬ Research Applications
This benchmark is designed for researchers working on:
- Multimodal Large Language Models
- Embodied AI Systems
- Robotic Intelligence
- Computer Vision
- Natural Language Processing
π Citation
If you use RoboBench in your research, please cite our paper:
@article{luo2025robobench,
title={Robobench: A Comprehensive Evaluation Benchmark for Multimodal Large Language Models as Embodied Brain},
author={Luo, Yulin and Fan, Chun-Kai and Dong, Menghang and Shi, Jiayu and Zhao, Mengdi and Zhang, Bo-Wen and Chi, Cheng and Liu, Jiaming and Dai, Gaole and Zhang, Rongyu and others},
journal={arXiv preprint arXiv:2510.17801},
year={2025}
}
π€ Contributing
We welcome contributions! Please see our Contributing Guidelines for more details.
π License
This dataset is released under the Creative Commons Attribution 4.0 International License.
π Links
- π Paper: arXiv:2510.17801
- π Project Page: https://robo-bench.github.io/
- π» GitHub: https://github.com/lyl750697268/RoboBench
Made with β€οΈ by the RoboBench Team