Porting HOST algorithm from Isaac Gym to Isaac Lab #164800
Replies: 3 comments
This comment was marked as spam.
This comment was marked as spam.
-
That’s an exciting project. Porting from Legged-Gym to Isaac Lab (Orbit) is a non-trivial but very rewarding effort, especially given the structured approach of your algorithm. I’ve worked a bit with both, so I’ll try to point you in the right direction based on your questions:
Isaac Orbit’s reward functions are modular and defined via RewardTerms. While multi-critic PPO isn’t built-in, you can simulate the structure by: Splitting the reward into stage-specific components using custom flags or environment variables. Logging these components separately (more on that below). Training a unified policy but conditioning the reward functions dynamically per stage. You could also explore using RL Games with Orbit, which supports custom reward shaping and has hooks you can leverage to implement multi-critic-like behavior.
Orbit provides CurriculumTerm support under RandomizationCfg, or you can implement curriculum via a callback that modifies environment parameters based on training step or agent success rate. Example strategy: Define curriculum variables in your environment config (like external forces). Use a custom CurriculumManager or implement the logic in your reset() and step() hooks to update difficulty. Also take a look at Orbit’s FrankaCubeStack curriculum example for inspiration. |
Beta Was this translation helpful? Give feedback.
-
🕒 Discussion Activity Reminder 🕒 This Discussion has been labeled as dormant by an automated system for having no activity in the last 60 days. Please consider one the following actions: 1️⃣ Close as Out of Date: If the topic is no longer relevant, close the Discussion as 2️⃣ Provide More Information: Share additional details or context — or let the community know if you've found a solution on your own. 3️⃣ Mark a Reply as Answer: If your question has been answered by a reply, mark the most helpful reply as the solution. Note: This dormant notification will only apply to Discussions with the Thank you for helping bring this Discussion to a resolution! 💬 |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I’m working with the team on porting a custom reinforcement learning algorithm from Isaac Gym (Legged-Gym) to Isaac Lab using Isaac Orbit, and I’d really appreciate any advice or guidance from the community.
The original implementation is based on the paper:
"Learning Humanoid Standing-Up Control Across Diverse Postures" by Tao Huang and partners.
The code is built upon Nvidia’s Legged-Gym library (using Isaac Gym), and defines a multi-stage standing-up behavior for a humanoid robot. The agent is trained with PPO and leverages custom design features like:
I want to recreate the same learning environment and behavior inside Isaac Lab, using the Orbit framework. Specifically:
What I'm looking for:
If you’ve worked on similar problems or have seen relevant examples, I’d love to hear from you. Thanks in advance for your time and any suggestions 🙏
Best regards,
Francesca
Beta Was this translation helpful? Give feedback.
All reactions