[Discussion] RoboGate: Adversarial Safety Benchmark for Pick-and-Place (68 scenarios, 50K+ failure dictionary) #508
liveplex-cpu
started this conversation in
Ideas
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
RoboGate — Adversarial Safety Benchmark for Industrial Pick-and-Place
Hi Isaac Lab-Arena team and community,
We would like to share RoboGate, an adversarial safety benchmark designed to complement existing Isaac Lab-Arena benchmarks like Lightwheel RoboFinals and RoboCasa Tasks.
What RoboGate Does
Focus: Pre-deployment safety validation — answering "Is this policy safe to deploy?" rather than "How well does this policy perform?"
5-Model VLA Leaderboard
All 4 VLA models — including NVIDIA's official GR00T N1.6 — score 0% on scenarios a scripted IK controller solves 100%.
How RoboGate Complements Lightwheel Benchmarks
The two approaches are complementary:
Together, they provide a complete picture: generalization breadth (Lightwheel) + safety depth (RoboGate).
Pull Request
We have an open PR integrating the benchmark into Isaac Lab-Arena:
--mockmode for CI/CD testing without GPUArenaEnvBuilderResources
We welcome feedback on how RoboGate can best integrate with the Arena ecosystem. Happy to adjust the benchmark design based on maintainer and community input.
— AgentAI Co., Ltd.
Beta Was this translation helpful? Give feedback.
All reactions