Datasets:
Update README.md
Browse files
README.md
CHANGED
|
@@ -23,7 +23,7 @@ configs:
|
|
| 23 |
## Dataset Description
|
| 24 |
|
| 25 |
<!-- Provide a longer summary of what this dataset is. -->
|
| 26 |
-
We introduce ViewSpatial-Bench
|
| 27 |
|
| 28 |
ViewSpatial-Bench addresses a critical gap: while VLMs excel at spatial reasoning from their own perspective, they struggle with perspective-taking—adopting another entity's spatial frame of reference—which is essential for embodied interaction and multi-agent collaboration.The figure below shows the construction pipeline and example demonstrations of our benchmark.
|
| 29 |
|
|
|
|
| 23 |
## Dataset Description
|
| 24 |
|
| 25 |
<!-- Provide a longer summary of what this dataset is. -->
|
| 26 |
+
We introduce **ViewSpatial-Bench**, a comprehensive benchmark with over 5,700 question-answer pairs across 1,000+ 3D scenes from ScanNet and MS-COCO validation sets. This benchmark evaluates VLMs' spatial localization capabilities from multiple perspectives, specifically testing both egocentric (camera) and allocentric (human subject) viewpoints across five distinct task types.
|
| 27 |
|
| 28 |
ViewSpatial-Bench addresses a critical gap: while VLMs excel at spatial reasoning from their own perspective, they struggle with perspective-taking—adopting another entity's spatial frame of reference—which is essential for embodied interaction and multi-agent collaboration.The figure below shows the construction pipeline and example demonstrations of our benchmark.
|
| 29 |
|