dataset_info: | |
features: | |
- name: images | |
sequence: image | |
- name: problem | |
dtype: string | |
- name: answer | |
dtype: string | |
splits: | |
- name: train | |
num_bytes: 167434471.0 | |
num_examples: 2000 | |
download_size: 166955903 | |
dataset_size: 167434471.0 | |
configs: | |
- config_name: default | |
data_files: | |
- split: train | |
path: data/train-* | |
This is the official release of the training data for paper **PAPO: Perception-Aware Policy Optimization for Multimodal Reasoning**. (arxiv.org/abs/2507.06448) | |
(Optional) This dataset can be used as the `val` split of the training dataset for PAPO. You may find the full training dataset at [PAPOGalaxy/PAPO_ViRL39K_train](https://huggingface.co/datasets/PAPOGalaxy/PAPO_ViRL39K_train). | |
# Data Source | |
## **Training** | |
- We adapt the multimodal benchmark [TIGER-Lab/ViRL39K](https://huggingface.co/datasets/TIGER-Lab/ViRL39K) to construct our PAPO training dataset. | |
## **Validation (Optional)** | |
- (Optional) We use the `test` set from [FanqingM/MMK12](https://huggingface.co/datasets/FanqingM/MMK12) for validation during training. | |
- Note that this is solely for monitoring. We do not pick checkpoints based on this in our paper. | |
# Dataset Structure | |
- **train:** training set consisting of **38870** multimodal reasoning samples | |
- **val:** validation set consisting of **2000** multimodal reasoning samples | |
# Data Fields | |
- **id:** data id | |
- data type: String | |
- **problem:** input question or statement | |
- - data type: String | |
- **images:** input image(s) | |
- data type: List | |
- **answer:** ground-truth answer | |
- - data type: String | |
# Usage | |
To use the full dataset with both `train` and `val` split, you may code as follows: | |
```python | |
# Train | |
train_dataset = load_dataset("PAPOGalaxy/PAPO_ViRL39K_train") | |
# Val | |
val_dataset = load_dataset("PAPOGalaxy/PAPO_MMK12_test") | |
``` |