Skip to content

Releases: takuseno/d3rlpy

Release version v0.23

08 Sep 04:16
Compare
Choose a tag to compare

Algorithm

  • Support Advantage-Weighted Regression (AWR)
  • n_frames option is added to all algorithms
    • n_frames option controls frame stacking for image observation
  • eval_results_ property is added to all algorithms
    • evaluation results can be retrieved from eval_results_ after training.

MDPDataset

  • prev_transition and next_transition properties are added to d3rlpy.dataset.Transition.
    • these properties are used for frame stacking and Monte-Carlo returns calculation at AWR.

Document

  • new tutorial page is added

Release version v0.22

28 Aug 01:22
Compare
Choose a tag to compare

Support ONNX export

Now, the trained policy can be exported as ONNX as well as TorchScript

cql.save_policy('policy.onnx', as_onnx=True)

Support more data augmentations

  • data augmentations for vector obsrevation
  • ColorJitter augmentation for image observation

Release version v0.2

10 Aug 23:54
Compare
Choose a tag to compare
  • support model-based algorithm
    • Model-based Offline Policy Optimization
  • support data augmentation (for image observation)
    • Data-reguralized Q-learning
  • a lot of improvements
    • more dataset statistics
    • more options to customize neural network architecture
    • optimize default learning rates
    • etc

First release!

31 Jul 15:40
Compare
Choose a tag to compare
  • online algorithms
    • Deep Q-Network (DQN)
    • Double DQN
    • Deep Deterministic Policy Gradients (DDPG)
    • Twin Delayed Deep Deterministic Policy Gradients (TD3)
    • Soft Actor-Critic (SAC)
  • data-driven algorithms
    • Batch-Constrained Q-leearning (BCQ)
    • Bootstrapping Error Accumulation Reduction (BEAR)
    • Conservative Q-Learning (CQL)
  • Q functions
    • mean
    • Quantile Regression
    • Implicit Quantile Network
    • Fully-parametrized Quantile Function (experimental)