Update README.md
Browse files
README.md
CHANGED
|
@@ -12,6 +12,8 @@ base_model:
|
|
| 12 |
|
| 13 |
<div align=center><img src="https://raw.githubusercontent.com/tongjingqi/Code2Logic/refs/heads/main/assets/evaluation_results_on_general_vision_benchmarks.png"></div>
|
| 14 |
|
|
|
|
|
|
|
| 15 |
It's also found that getting trained on 5k samples from our GameQA dataset can lead to better results than on [multimodal-open-r1-8k-verified](https://huggingface.co/datasets/lmms-lab/multimodal-open-r1-8k-verified).
|
| 16 |
|
| 17 |
<div align=center><img src="https://raw.githubusercontent.com/tongjingqi/Code2Logic/refs/heads/main/assets/GameQA_generalizes_better.png"></div>
|
|
|
|
| 12 |
|
| 13 |
<div align=center><img src="https://raw.githubusercontent.com/tongjingqi/Code2Logic/refs/heads/main/assets/evaluation_results_on_general_vision_benchmarks.png"></div>
|
| 14 |
|
| 15 |
+
***(The inference and evaluation configurations were unified across both the original open-source models and our trained models.)***
|
| 16 |
+
|
| 17 |
It's also found that getting trained on 5k samples from our GameQA dataset can lead to better results than on [multimodal-open-r1-8k-verified](https://huggingface.co/datasets/lmms-lab/multimodal-open-r1-8k-verified).
|
| 18 |
|
| 19 |
<div align=center><img src="https://raw.githubusercontent.com/tongjingqi/Code2Logic/refs/heads/main/assets/GameQA_generalizes_better.png"></div>
|