The format is based on Keep a Changelog.
Summary of the most relevant features:
- RL algorithm implementations in NVIDIA Warp
- Differentiate between environment observations and states (also known as privileged observation)
- Support for MuJoCo Playground and ManiSkill environments
- Implement RL algorithms in NVIDIA Warp
- Add loader and wrapper for MuJoCo Playground environments
- Add wrapper for ManiSkill environments
- Add Tabular model instantiator (epsilon-greedy variant)
- Add
clip_mean_actionsparameter to Gaussian and Multivariate Gaussian models - Add
render_intervaloption to trainers to specify the rendering interval for the environments - Add
compute_space_limitsspace utility to get Gymnasium spaces' limits - Add
ScopedTimerutils to measure code execution time - Add
SummaryWriterimplementation to log data to TensorBoard without relying on third-party libraries - Log agent inference and algorithm update, and environment steeping time to TensorBoard
- Update minimum supported Python version to 3.10
- Drop support for PyTorch versions prior to 1.11 (the previous supported version was 1.10)
- Call observation/state preprocessors once when computing the actions during training
- Refactor the library to differentiate between environment observations and states (also known as privileged observation)
- Implement agent/multi-agent and trainer configurations using Python Data Classes
- Unify the different learning rate settings under the
learning_rateconfiguration - Rename
lambdatogae_lambda - Remove the
clip_predicted_valuesredundant configuration by checking forvalue_clip > 0 - Remove specific exploration noise settings (
initial_scale,final_scaleandtimesteps) in favor of generic scheduling functions
- Unify the different learning rate settings under the
- Update tabular model definition to operate in any number of parallel environments
- Refactor multi-agent environment wrappers to support homogeneous and heterogeneous states spaces
- Add entropy loss to the policy loss for on-policy agents/mulit-agents in JAX
- Fix time limits handling for termination and truncation signals
- Fix the randomness of the environments by seeding right after initialization (on the first reset)
- Remove NumPy backend for JAX implementation
- Remove checkpoints/models migration support from other RL libraries
- Remove support for Isaac Gym and Omniverse Isaac Gym environments (deprecated in favor of Isaac Lab)
- Remove support for Brax and DeepMind environments (in favor of MuJoCo Playground environments)
- Remove support for Bi-DexHands and robosuite environments
- Remove Isaac Gym (web viewer, inverse kinematic) and Omniverse Isaac Gym (local environment instance, inverse kinematic) utils
- Update the GitHub Actions workflows for testing and coverage
- Update minimum supported Python version to 3.8 and minimum dependencies versions
- Fix environment wrapper issues with spaces utilities's keyword-only arguments (introduced in previous version)
- Fix noise device definition in runner implementations
- Add Multi-Categorical model instantiator
- Add
one_hot_encodingfunction to model instantiators to one-hot encodeDiscreteandMultiDiscretetensorized spaces - Allow
Nonetype spaces and samples/values in spaces utilities and define keyword-only arguments
- Cast model instantiator's
initial_log_stdparameter tofloatin PyTorch - Fix common property overwriting (e.g.
clip_actions) in shared models composed of different mixin types
- Force the use of the device local to process in distributed runs in JAX
- Update runner implementation to parse noises definitions for off-policy agents
- Utilities to operate on Gymnasium spaces (
Box,Discrete,MultiDiscrete,TupleandDict) parse_devicestatic method in ML framework configuration (used in library components to set up the device)- Model instantiator support for different shared model structures in PyTorch
- Support for automatic mixed precision training in PyTorch
init_state_dictmethod to initialize model's lazy modules in PyTorch- Model instantiators
fixed_log_stdparameter to define immutable log standard deviations - Define the
stochastic_evaluationtrainer config to allow the use of the actions returned by the agent's model as-is instead of deterministic actions (mean-actions in Gaussian-based models) during evaluation. Make the return of deterministic actions the default behavior.
- Call agent's
pre_interactionmethod during evaluation - Use spaces utilities to process states, observations and actions for all the library components
- Update model instantiators definitions to process supported fundamental and composite Gymnasium spaces
- Make flattened tensor storage in memory the default option (revert changed introduced in version 1.3.0)
- Drop support for PyTorch versions prior to 1.10 (the previous supported version was 1.9)
- Update KL Adaptive learning rate scheduler implementation to match Optax's behavior in JAX
- Update AMP agent to use the environment's terminated and truncated data, and the KL Adaptive learning rate scheduler
- Update runner implementations to support definition of arbitrary agents and their models
- Speed up PyTorch implementation:
- Disable argument checking when instantiating distributions
- Replace PyTorch's
BatchSamplerby Python slice when sampling data from memory
- Format code using Black code formatter (it's ugly, yes, but it does its job)
- Move the batch sampling inside gradient step loop for DQN, DDQN, DDPG (RNN), TD3 (RNN), SAC and SAC (RNN)
- Model state dictionary initialization for composite Gymnasium spaces in JAX
- Add missing
reductionparameter to Gaussian model instantiator - Optax's learning rate schedulers integration in JAX implementation
- Isaac Lab wrapper's multi-agent state retrieval with gymnasium 1.0
- Treat truncation signal when computing 'done' (environment reset)
- Remove OpenAI Gym (
gym) from dependencies and source code. skrl continues to support gym environments, it is just not installed as part of the library. If it is needed, it needs to be installed manually. Any gym-based environment wrapper must use theconvert_gym_spacespace utility to operate
- Distributed multi-GPU and multi-node learning (JAX implementation)
- Utilities to start multiple processes from a single program invocation for distributed learning using JAX
- Model instantiators
return_sourceparameter to get the source class definition used to instantiate the models Runnerutility to run training/evaluation workflows in a few lines of code- Wrapper for Isaac Lab multi-agent environments
- Wrapper for Google Brax environments
- Move the KL reduction from the PyTorch
KLAdaptiveLRclass to each agent that uses it in distributed runs - Move the PyTorch distributed initialization from the agent base class to the ML framework configuration
- Upgrade model instantiator implementations to support CNN layers and complex network definitions, and implement them using dynamic execution of Python code
- Update Isaac Lab environment loader argument parser options to match Isaac Lab version
- Allow to store tensors/arrays with their original dimensions in memory and make it the default option
- Decouple the observation and state spaces in single and multi-agent environment wrappers and add the
statemethod to get the state of the environment - Simplify multi-agent environment wrapper API by removing shared space properties and methods
- Catch TensorBoard summary iterator exceptions in
TensorboardFileIteratorpostprocessing utils - Fix automatic wrapper detection issue (introduced in previous version) for Isaac Gym (previews), DeepMind and vectorized Gymnasium environments
- Fix vectorized/parallel environments
resetmethod return values when called more than once - Fix IPPO and MAPPO
actmethod return values when JAX-NumPy backend is enabled
- Define the
environment_infotrainer config to log environment info (PyTorch implementation) - Add support to automatically compute the write and checkpoint intervals and make it the default option
- Single forward-pass in shared models
- Distributed multi-GPU and multi-node learning (PyTorch implementation)
- Update Orbit-related source code and docs to Isaac Lab
- Move the batch sampling inside gradient step loop for DDPG and TD3
- Perform JAX computation on the selected device
MultiCategoricalMixinto operateMultiDiscreteaction spaces
- Rename the
ManualTrainertoStepTrainer - Output training/evaluation progress messages to system's stdout
- Get single observation/action spaces for vectorized environments
- Update Isaac Orbit environment wrapper
Transition from pre-release versions (1.0.0-rc.1 and1.0.0-rc.2) to a stable version.
This release also announces the publication of the skrl paper in the Journal of Machine Learning Research (JMLR): https://www.jmlr.org/papers/v24/23-0112.html
Summary of the most relevant features:
- RL algorithm implementations in JAX
- New documentation theme and structure
- Multi-agent Reinforcement Learning (MARL)
- Get truncation from
time_outsinfo in Isaac Gym, Isaac Orbit and Omniverse Isaac Gym environments - Time-limit (truncation) bootstrapping in on-policy actor-critic agents
- Model instantiators
initial_log_stdparameter to set the log standard deviation's initial value
- Structure environment loaders and wrappers file hierarchy coherently.
Import statements now follow the next convention:
- Wrappers (e.g.):
from skrl.envs.wrappers.torch import wrap_envfrom skrl.envs.wrappers.jax import wrap_env
- Loaders (e.g.):
from skrl.envs.loaders.torch import load_omniverse_isaacgym_envfrom skrl.envs.loaders.jax import load_omniverse_isaacgym_env
- Wrappers (e.g.):
- Drop support for PyTorch versions prior to 1.9 (the previous supported version was 1.8)
- Implement RL algorithms in JAX (Flax/Optax)
- RPO agent
- IPPO and MAPPO multi-agent
- Multi-agent base class
- Bi-DexHands environment loader
- Wrapper for Bi-DexHands environments
- Wrapper for PettingZoo environments
- Parameters
num_envs,headlessandcli_argsfor configuring Isaac Gym, Isaac Orbit and Omniverse Isaac Gym environments when they are loaded
- Migrate to
pyproject.tomlPython package development - Define ML framework dependencies as optional dependencies in the library installer
- Move agent implementations with recurrent models to a separate file
- Allow closing the environment at the end of execution instead of after training/evaluation
- Documentation theme from sphinx_rtd_theme to furo
- Update documentation structure and examples
- Compatibility for Isaac Sim or OmniIsaacGymEnvs (2022.2.0 or earlier)
- Disable PyTorch gradient computation during the environment stepping
- Get categorical models' entropy
- Typo in
KLAdaptiveLRlearning rate scheduler (Keep the old name for compatibility with the examples of previous versions. The old name will be removed in future releases)
- Update loader and utils for OmniIsaacGymEnvs 2022.2.1.0
- Update Omniverse Isaac Gym real-world examples
- TensorBoard writer instantiation when
write_intervalis zero
- Isaac Orbit environment loader
- Wrap an Isaac Orbit environment
- Gaussian-Deterministic shared model instantiator
- Utility for downloading models from Hugging Face Hub
- Initialization of agent components if they have not been defined
- Manual trainer
train/evalmethod default arguments
- Support for Farama Gymnasium interface
- Wrapper for robosuite environments
- Weights & Biases integration
- Set the running mode (training or evaluation) of the agents
- Allow clipping the gradient norm for DDPG, TD3 and SAC agents
- Initialize model biases
- Add RNN (RNN, LSTM, GRU and any other variant) support for A2C, DDPG, PPO, SAC, TD3 and TRPO agents
- Allow disabling training/evaluation progressbar
- Farama Shimmy and robosuite examples
- KUKA LBR iiwa real-world example
- Forward model inputs as a Python dictionary
- Returns a Python dictionary with extra output values in model calls
- Adopt the implementation of
terminatedandtruncatedoverdonefor all environments
- Omniverse Isaac Gym simulation speed for the Franka Emika real-world example
- Call agents' method
record_transitioninstead of parent method to allow storing samples in memories during evaluation - Move TRPO policy optimization out of the value optimization loop
- Access to the categorical model distribution
- Call reset only once for Gym/Gymnasium vectorized environments
- Deprecated method
startin trainers
- AMP agent for physics-based character animation
- Manual trainer
- Gaussian model mixin
- Support for creating shared models
- Parameter
roleto model methods - Wrapper compatibility with the new OpenAI Gym environment API
- Internal library colored logger
- Migrate checkpoints/models from other RL libraries to skrl models/agents
- Configuration parameter
store_separatelyto agent configuration dict - Save/load agent modules (models, optimizers, preprocessors)
- Set random seed and configure deterministic behavior for reproducibility
- Benchmark results for Isaac Gym and Omniverse Isaac Gym on the GitHub discussion page
- Franka Emika real-world example
- Models implementation as Python mixin
- Multivariate Gaussian model (
GaussianModeluntil 0.7.0) toMultivariateGaussianMixin - Trainer's
cfgparameter position and default values - Show training/evaluation display progress using
tqdm - Update Isaac Gym and Omniverse Isaac Gym examples
- Missing recursive arguments during model weights initialization
- Tensor dimension when computing preprocessor parallel variance
- Models' clip tensors dtype to
float32
- Parameter
inferencefrom model methods - Configuration parameter
checkpoint_policy_onlyfrom agent configuration dict
- A2C agent
- Isaac Gym (preview 4) environment loader
- Wrap an Isaac Gym (preview 4) environment
- Support for OpenAI Gym vectorized environments
- Running standard scaler for input preprocessing
- Installation from PyPI (
pip install skrl)
- Omniverse Isaac Gym environment loader
- Wrap an Omniverse Isaac Gym environment
- Save best models during training
- TRPO agent
- Wrapper for DeepMind environments
- KL Adaptive learning rate scheduler
- Handle
gym.spaces.Dictobservation spaces (OpenAI Gym and DeepMind environments) - Forward environment info to agent
record_transitionmethod - Expose and document the random seeding mechanism
- Define rewards shaping function in agents' config
- Define learning rate scheduler in agents' config
- Improve agent's algorithm description in documentation (PPO and TRPO at the moment)
- Compute the Generalized Advantage Estimation (GAE) in agent
_updatemethod - Move noises definition to
resourcesfolder - Update the Isaac Gym examples
compute_functionsfor computing the GAE from memory base class
- Examples of all Isaac Gym environments (preview 3)
- TensorBoard file iterator for data post-processing
- Init and evaluate agents in ParallelTrainer
- CEM, SARSA and Q-learning agents
- Tabular model
- Parallel training using multiprocessing
- Isaac Gym utilities
- Initialize agents in a separate method
- Change the name of the
networksargument tomodels
- Reset environments after post-processing
- DQN and DDQN agents
- Export memory to files
- Postprocessing utility to iterate over memory files
- Model instantiator utility to allow fast development
- More examples and contents in the documentation
- Clip actions using the whole space's limits
- First official release