RoboGround: Robotic Manipulation with Grounded Vision-Language Priors Paper • 2504.21530 • Published Apr 30
CronusVLA: Transferring Latent Motion Across Time for Multi-Frame Prediction in Manipulation Paper • 2506.19816 • Published Jun 24
GENMANIP: LLM-driven Simulation for Generalizable Instruction-Following Manipulation Paper • 2506.10966 • Published Jun 12
MM-ACT: Learn from Multimodal Parallel Generation to Act Paper • 2512.00975 • Published 27 days ago • 6
ARM-Thinker: Reinforcing Multimodal Generative Reward Models with Agentic Tool Use and Visual Reasoning Paper • 2512.05111 • Published 23 days ago • 45
MM-ACT: Learn from Multimodal Parallel Generation to Act Paper • 2512.00975 • Published 27 days ago • 6
InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policy Paper • 2510.13778 • Published Oct 15 • 16
InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policy Paper • 2510.13778 • Published Oct 15 • 16
InternSVG: Towards Unified SVG Tasks with Multimodal Large Language Models Paper • 2510.11341 • Published Oct 13 • 34
MCPMark: A Benchmark for Stress-Testing Realistic and Comprehensive MCP Use Paper • 2509.24002 • Published Sep 28 • 173
F1: A Vision-Language-Action Model Bridging Understanding and Generation to Actions Paper • 2509.06951 • Published Sep 8 • 32
The Imitation Game: Turing Machine Imitator is Length Generalizable Reasoner Paper • 2507.13332 • Published Jul 17 • 48
SmolVLA: A Vision-Language-Action Model for Affordable and Efficient Robotics Paper • 2506.01844 • Published Jun 2 • 147
MM-Eureka: Exploring Visual Aha Moment with Rule-based Large-scale Reinforcement Learning Paper • 2503.07365 • Published Mar 10 • 61