This is the Linux app named RLax whose latest release can be downloaded as RLax0.1.8sourcecode.tar.gz. It can be run online in the free hosting provider OnWorks for workstations.
Download and run online this app named RLax with OnWorks for free.
Befolgen Sie diese Anweisungen, um diese App auszuführen:
- 1. Diese Anwendung auf Ihren PC heruntergeladen.
- 2. Geben Sie in unserem Dateimanager https://www.onworks.net/myfiles.php?username=XXXXX den gewünschten Benutzernamen ein.
- 3. Laden Sie diese Anwendung in einem solchen Dateimanager hoch.
- 4. Starten Sie den OnWorks Linux-Online- oder Windows-Online-Emulator oder den MACOS-Online-Emulator von dieser Website.
- 5. Rufen Sie vom gerade gestarteten OnWorks Linux-Betriebssystem aus unseren Dateimanager https://www.onworks.net/myfiles.php?username=XXXXX mit dem gewünschten Benutzernamen auf.
- 6. Laden Sie die Anwendung herunter, installieren Sie sie und führen Sie sie aus.
SCREENSHOTS
Ad
RLax
BESCHREIBUNG
RLax (pronounced “relax”) is a JAX-based library developed by Google DeepMind that provides reusable mathematical building blocks for constructing reinforcement learning (RL) agents. Rather than implementing full algorithms, RLax focuses on the core functional operations that underpin RL methods—such as computing value functions, returns, policy gradients, and loss terms—allowing researchers to flexibly assemble their own agents. It supports both on-policy and off-policy learning, as well as value-based, policy-based, and model-based approaches. RLax is fully JIT-compilable with JAX, enabling high-performance execution across CPU, GPU, and TPU backends. The library implements tools for Bellman equations, return distributions, general value functions, and policy optimization in both continuous and discrete action spaces. It integrates seamlessly with DeepMind’s Haiku (for neural network definition) and Optax (for optimization), making it a key component in modular RL pipelines.
Eigenschaften
- Modular reinforcement learning primitives (values, returns, and policies)
- JAX-optimized for GPU/TPU acceleration and automatic differentiation
- Supports on-policy and off-policy learning paradigms
- Implements distributional value functions and general value functions
- Integrates with Haiku and Optax for neural network and optimization pipelines
- Comprehensive testing and examples for reproducibility and educational use
Programmiersprache
Python, Unix-Shell
Kategorien
This is an application that can also be fetched from https://sourceforge.net/projects/rlax.mirror/. It has been hosted in OnWorks in order to be run online in an easiest way from one of our free Operative Systems.