SLM Lab
v4.1.1
v4.1.1
  • SLM Lab
  • 🖥Setup
    • Installation
    • Quick Start
  • 🚀Using SLM Lab
    • Lab Command
    • Lab Organization
    • Train and Enjoy: REINFORCE CartPole
    • Agent Spec: DDQN+PER on LunarLander
    • Env Spec: A2C on Pong
    • GPU Usage: PPO on Pong
    • Parallelizing Training: Async SAC on Humanoid
    • Experiment and Search Spec: PPO on Breakout
    • Run Benchmark: A2C on Atari Games
    • Meta Spec: High Level Specifications
    • Post-Hoc Analysis
    • TensorBoard: Visualizing Models and Actions
    • Using SLM Lab In Your Project
  • 📈Analyzing Results
    • Data Locations
    • Graphs and Data
    • Performance Metrics
  • 🥇Benchmark Results
    • Public Benchmark Data
    • Discrete Environment Benchmark
    • Continuous Environment Benchmark
    • Atari Environment Benchmark
    • RL GIFs
  • 🔧Development
    • Modular Design
      • Algorithm Taxonomy
      • Class Inheritance: A2C > PPO
    • Algorithm
      • DQN
      • REINFORCE
      • Actor Critic
    • Memory
      • Replay
      • PrioritizedReplay
      • OnPolicyReplay
      • OnPolicyBatchReplay
    • Net
      • MLP
      • CNN
      • RNN
    • Profiling SLM Lab
  • 📖Publications and Talks
    • Book: Foundations of Deep Reinforcement Learning
    • Talks and Presentations
  • 🤓Resources
    • Deep RL Resources
    • Contributing
    • Motivation
    • Help
    • Contact
Powered by GitBook
On this page
  • Public Data
  • Benchmark Information
  • Hardware
  • Reproducibility
  • Environments
  • Terminology

Was this helpful?

  1. 🥇Benchmark Results

Public Benchmark Data

PreviousPerformance MetricsNextDiscrete Environment Benchmark

Last updated 5 years ago

Was this helpful?

Public Data

SLM Lab provides a set of benchmark results that are periodically updated with new feature releases. All the result data is and made .

The data can be downloaded and unzipped into SLM Lab's data/ folder and rerun in .

Benchmark Information

Hardware

For reference, the image based environment benchmarks are run on AWS GPU box p2.16xlarge, and the non-image based environments are run on AWS CPU box m5a.24xlarge.

Reproducibility

The benchmark tables in this page show the Trial level final_return_ma from SLM Lab. This is final value of the 100-ckpt moving average of the return (total rewards) from evaluation. Each Trial is ran with 4 Sessions with different random seeds, and their final_return_ma are averaged on the Trial level.

The specs for these are contained in the folder, descriptively named {algorithm}_{environment}.json. They can be exactly reproduced as described in .

Environments

SLM Lab's benchmark includes environments from the following offerings:

  • offers a wrapper for the

Terminology

Deep RL algorithms use a lot of abbreviations. Here's a list to help us navigate:

  • A2C (GAE): Advantage Actor-Critic with GAE as advantage estimation

  • A2C (n-step): Advantage Actor-Critic with n-step return as advantage estimation

  • A3C: Asynchronous Advantage Actor-Critic

  • CER: Combined Experience Replay

  • DDQN: Double Deep Q-Network

  • Async: Asynchronous

  • DQN: Deep Q-Network

  • GAE: Generalized Advantage Estimation

  • PER: Prioritized Experience Replay

  • PPO: Proximal Policy Optimization

  • SAC: Soft Actor-Critic

  • SIL: Self Imitation Learning

Read on to see the benchmark result tables and plots.

📂
📌
uploaded from a Pull Request
public on Dropbox
enjoy mode
slm_lab/spec/benchmark
OpenAI gym default environments
OpenAI gym Atari environments
Atari Learning Environment (ALE)
OpenAI Roboschool
Unity ML Agents
Lab Organization