| --- |
| pretty_name: VisPhyBench |
| license: mit |
| language: |
| - en |
| --- |
| |
| # VisPhyBench |
|
|
| To evaluate how well models reconstruct appearance and reproduce physically plausible motion, we introduce VisPhyBench, a unified evaluation protocol comprising 209 scenes derived from 108 physical templates that assesses physical understanding through the lens of code-driven resimulation in both 2D and 3D scenes, integrating metrics from different aspects. Each scene is also annotated with a coarse difficulty label (easy/medium/hard). |
|
|
| # Dataset Details |
|
|
| - **Created by:** Jiarong Liang |
| - **Language(s) (NLP):** English |
| - **License:** MIT |
| - **Repository:** `https://github.com/TIGER-AI-Lab/VisPhyWorld` |
|
|
| # Uses |
|
|
| The dataset is used to evaluate how well models reconstruct appearance and reproduce physically plausible motion. |
|
|
| # Dataset Structure |
|
|
| The difficulty of the two sets of VisPhyBench splits, **sub (209)** and **test (49)**, are as follows: **sub** is **114/67/28** (**54.5%/32.1%/13.4%**), and **test** is **29/17/3** (**59.2%/34.7%/6.1%**) (Easy/Medium/Hard). |
|
|
| ## What each sample contains |
|
|
| VisPhyBench is provided as two splits: |
|
|
| - `sub`: a larger split intended for evaluation and analysis. |
| - `test`: a smaller split subsampled from `sub` for quick sanity checks. |
|
|
| For each sample, we provide: |
|
|
| 1. **A short video** of a synthetic physical scene. |
| 2. **A detection JSON** (per sample) that describes the scene in the first frame. |
| 3. **A difficulty label** (easy/medium/hard) derived from the mean of eight annotators’ 1–5 ratings. |
|
|
| ## Detection JSON format |
|
|
| Each detection JSON includes: |
|
|
| - `image_size`: the image width/height. |
| - `coordinate_system`: conventions for coordinates (e.g., origin and axis directions). |
| - `objects`: a list of detected objects. Each object includes: |
| - `id`: unique identifier. |
| - `category`: coarse geometry category. |
| - `color_rgb`: RGB color triplet. |
| - `position`: object position (e.g., center coordinates). |
| - `bbox`: bounding box coordinates and size. |
| - `size`: coarse size fields (e.g., radius/length/thickness). |
|
|
| These fields specify object locations and attributes precisely, which helps an LLM initialize objects correctly when generating executable simulation code. |
|
|
| # BibTeX |
|
|
| ```bibtex |
| @misc{visphybench2026, |
| title = {VisPhyBench}, |
| author = {Liang, Jiarong and Ku, Max and Hui, Ka-Hei and Nie, Ping and Chen, Wenhu}, |
| howpublished = {GitHub repository}, |
| year = {2026}, |
| url = {https://github.com/TIGER-AI-Lab/VisPhyWorld} |
| } |
| ``` |
|
|
|
|