Learning to Walk in Minutes Using Massively Parallel Deep Reinforcement Learning (ANYmal)
Product* ⭐ 10/2022: [Discovering faster matrix multiplication algorithms with reinforcement learning (AlphaTensor)](https://www.nature.com/articles/s41586-022%20-05172-4)
Capabilities6 decomposed
massively-parallel distributed reinforcement learning training
Medium confidenceTrains quadruped locomotion policies using distributed deep RL across thousands of parallel simulation environments running synchronously on GPU clusters. The system uses PPO (Proximal Policy Optimization) with vectorized environment sampling, enabling wall-clock training times measured in minutes rather than hours or days. Implements gradient accumulation and asynchronous parameter updates across distributed workers to maintain training stability while maximizing throughput.
Achieves training convergence in minutes through extreme parallelization (thousands of synchronous environments) combined with PPO's sample-efficient policy gradient updates, enabled by vectorized GPU-accelerated physics simulation rather than sequential rollouts
Trains quadruped policies 100-1000x faster than traditional sequential RL by leveraging GPU-vectorized simulation and distributed PPO, compared to CPU-based or single-environment approaches
domain randomization for sim-to-real transfer
Medium confidenceAutomatically varies simulation parameters (friction, mass, inertia, actuator delays, sensor noise) during training to create a distribution of physics models that the learned policy must generalize across. The system samples randomization parameters from predefined ranges at each episode reset, forcing the policy to learn robust behaviors invariant to model mismatch. This approach reduces the need for manual real-world tuning by training policies that work across a wide range of physical conditions.
Applies curriculum-style domain randomization across thousands of parallel environments, sampling new randomization parameters per episode to create an implicit ensemble of physics models that the policy must simultaneously adapt to
Achieves real-world transfer without manual tuning by training against a distribution of simulated physics, compared to single-model simulation training that typically requires extensive real-world fine-tuning
gpu-accelerated vectorized physics simulation
Medium confidenceExecutes thousands of parallel robot simulations simultaneously on GPU hardware using a vectorized physics engine (Isaac Gym), where each environment step is computed in parallel across CUDA threads. The system batches environment state, action, and physics computations into tensor operations, eliminating the sequential bottleneck of traditional CPU-based simulators. This enables sampling millions of environment transitions per second, critical for training deep RL policies with massive batch sizes.
Implements fully vectorized physics simulation on GPU where all 4000+ environments execute in parallel as tensor operations, rather than sequential CPU simulation loops, achieving 1000x throughput improvement
Samples transitions 100-1000x faster than CPU-based simulators (PyBullet, MuJoCo) by executing all environments as batched GPU tensor operations rather than sequential simulation steps
end-to-end neural network policy learning for quadruped locomotion
Medium confidenceLearns a neural network policy that maps raw sensor observations (joint angles, velocities, IMU readings, contact forces) directly to motor commands (joint torques) using PPO with a multi-layer perceptron architecture. The policy is trained end-to-end via policy gradient optimization without hand-crafted features or inverse kinematics, discovering locomotion gaits emergently from reward signals. The learned policy encodes implicit knowledge of robot dynamics, balance, and gait coordination in its weights.
Learns locomotion policies entirely from raw sensor inputs to motor outputs via PPO without any hand-crafted features, inverse kinematics, or gait primitives, discovering natural gaits emergently through distributed RL training
Eliminates hand-coded controllers and gait libraries by learning end-to-end policies that adapt to new tasks and terrains, compared to traditional inverse kinematics and trajectory planning approaches
reward shaping and curriculum learning for complex locomotion tasks
Medium confidenceStructures reward functions to guide policy learning toward desired locomotion behaviors (e.g., forward velocity, energy efficiency, stability) and progressively increases task difficulty during training. The system decomposes complex objectives into reward components (velocity bonus, energy penalty, stability bonus) that are weighted and combined. Curriculum learning gradually increases terrain difficulty, speed targets, or disturbance magnitude as the policy improves, preventing early convergence to suboptimal solutions.
Combines multi-component reward shaping with progressive curriculum learning, where task difficulty increases automatically as policy performance improves, enabling stable training toward complex locomotion objectives
Guides RL training toward natural, energy-efficient gaits by decomposing objectives into weighted reward components and progressively increasing difficulty, compared to sparse reward or single-objective approaches
real-time policy inference on robot hardware
Medium confidenceDeploys trained neural network policies directly on robot onboard compute (CPU or GPU) for real-time motor control at 50-100 Hz control frequencies. The system quantizes and optimizes the policy network for inference latency, enabling sub-10ms inference times suitable for closed-loop control. Policies run autonomously without cloud connectivity, using only local sensor readings to generate motor commands.
Optimizes trained policies for sub-10ms inference on robot onboard compute through quantization and model optimization, enabling fully autonomous real-time control without cloud connectivity
Enables autonomous real-time control by deploying optimized policies directly on robot hardware, compared to cloud-based inference which introduces latency and connectivity dependencies
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with Learning to Walk in Minutes Using Massively Parallel Deep Reinforcement Learning (ANYmal), ranked by overlap. Discovered automatically through the match graph.
Outracing champion Gran Turismo drivers with deep reinforcement learning (Sophy)
* ⭐ 02/2022: [Magnetic control of tokamak plasmas through deep reinforcement learning](https://www.nature.com/articles/s41586-021-04301-9%E2%80%A6)
ray
Ray provides a simple, universal API for building distributed applications.
Ray
Distributed AI framework — Ray Train, Serve, Data, Tune for scaling ML workloads.
Learning robust perceptive locomotion for quadrupedal robots in the wild
* ⭐ 02/2022: [BC-Z: Zero-Shot Task Generalization with Robotic Imitation Learning](https://proceedings.mlr.press/v164/jang22a.html)
RT-1: Robotics Transformer for Real-World Control at Scale (RT-1)
## Historical Papers <a name="history"></a>
AReaL
The RL Bridge for LLM-based Agent Applications. Made Simple & Flexible.
Best For
- ✓robotics teams with access to large GPU clusters (8+ GPUs minimum)
- ✓researchers studying RL scalability and sample efficiency
- ✓organizations deploying quadruped robots requiring rapid policy adaptation
- ✓robotics teams deploying sim-trained policies to real quadrupeds
- ✓researchers studying robustness and generalization in RL
- ✓organizations with limited real-world robot access for validation
- ✓RL researchers training on large GPU clusters
- ✓robotics teams with access to high-end GPUs (A100, H100, RTX 6000)
Known Limitations
- ⚠Requires massive computational resources (thousands of parallel environments) — not feasible on single-GPU setups
- ⚠Training convergence depends heavily on hyperparameter tuning for specific robot morphologies
- ⚠Sim-to-real gap still requires domain randomization and careful reward shaping
- ⚠Limited to continuous control tasks with differentiable physics simulators
- ⚠Requires careful tuning of randomization ranges — too narrow fails to transfer, too wide prevents convergence
- ⚠Cannot handle systematic sim-to-real gaps (e.g., unmodeled contact dynamics, cable routing)
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
* ⭐ 10/2022: [Discovering faster matrix multiplication algorithms with reinforcement learning (AlphaTensor)](https://www.nature.com/articles/s41586-022%20-05172-4)
Categories
Alternatives to Learning to Walk in Minutes Using Massively Parallel Deep Reinforcement Learning (ANYmal)
Are you the builder of Learning to Walk in Minutes Using Massively Parallel Deep Reinforcement Learning (ANYmal)?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →