RANGER: A Monocular Zero-Shot Semantic Navigation Framework through Contextual Adaptation Paper • 2512.24212 • Published Dec 30, 2025
EgoActor: Grounding Task Planning into Spatial-aware Egocentric Actions for Humanoid Robots via Visual-Language Models Paper • 2602.04515 • Published 1 day ago • 32
VLS: Steering Pretrained Robot Policies via Vision-Language Models Paper • 2602.03973 • Published 2 days ago • 18
Likelihood-Based Reward Designs for General LLM Reasoning Paper • 2602.03979 • Published 2 days ago • 7
EgoActor: Grounding Task Planning into Spatial-aware Egocentric Actions for Humanoid Robots via Visual-Language Models Paper • 2602.04515 • Published 1 day ago • 32
EgoActor: Grounding Task Planning into Spatial-aware Egocentric Actions for Humanoid Robots via Visual-Language Models Paper • 2602.04515 • Published 1 day ago • 32
Self-Hinting Language Models Enhance Reinforcement Learning Paper • 2602.03143 • Published 3 days ago • 19
VIOLA: Towards Video In-Context Learning with Minimal Annotations Paper • 2601.15549 • Published 15 days ago • 4
Cosmos Policy: Fine-Tuning Video Models for Visuomotor Control and Planning Paper • 2601.16163 • Published 14 days ago • 13
PROGRESSLM: Towards Progress Reasoning in Vision-Language Models Paper • 2601.15224 • Published 15 days ago • 12
Terminal-Bench: Benchmarking Agents on Hard, Realistic Tasks in Command Line Interfaces Paper • 2601.11868 • Published 20 days ago • 32
EvoCUA: Evolving Computer Use Agents via Learning from Scalable Synthetic Experience Paper • 2601.15876 • Published 15 days ago • 89
SOP: A Scalable Online Post-Training System for Vision-Language-Action Models Paper • 2601.03044 • Published about 1 month ago • 28
Rethinking Video Generation Model for the Embodied World Paper • 2601.15282 • Published 15 days ago • 42
ShowUI-π: Flow-based Generative Models as GUI Dexterous Hands Paper • 2512.24965 • Published Dec 31, 2025 • 42