SLM Lab
v4.1.1
v4.1.1
  • SLM Lab
  • 🖥Setup
    • Installation
    • Quick Start
  • 🚀Using SLM Lab
    • Lab Command
    • Lab Organization
    • Train and Enjoy: REINFORCE CartPole
    • Agent Spec: DDQN+PER on LunarLander
    • Env Spec: A2C on Pong
    • GPU Usage: PPO on Pong
    • Parallelizing Training: Async SAC on Humanoid
    • Experiment and Search Spec: PPO on Breakout
    • Run Benchmark: A2C on Atari Games
    • Meta Spec: High Level Specifications
    • Post-Hoc Analysis
    • TensorBoard: Visualizing Models and Actions
    • Using SLM Lab In Your Project
  • 📈Analyzing Results
    • Data Locations
    • Graphs and Data
    • Performance Metrics
  • 🥇Benchmark Results
    • Public Benchmark Data
    • Discrete Environment Benchmark
    • Continuous Environment Benchmark
    • Atari Environment Benchmark
    • RL GIFs
  • 🔧Development
    • Modular Design
      • Algorithm Taxonomy
      • Class Inheritance: A2C > PPO
    • Algorithm
      • DQN
      • REINFORCE
      • Actor Critic
    • Memory
      • Replay
      • PrioritizedReplay
      • OnPolicyReplay
      • OnPolicyBatchReplay
    • Net
      • MLP
      • CNN
      • RNN
    • Profiling SLM Lab
  • 📖Publications and Talks
    • Book: Foundations of Deep Reinforcement Learning
    • Talks and Presentations
  • 🤓Resources
    • Deep RL Resources
    • Contributing
    • Motivation
    • Help
    • Contact
Powered by GitBook
On this page
  • Continuous Environment Benchmark Result
  • Continuous Environment Benchmark Result Plots

Was this helpful?

  1. 🥇Benchmark Results

Continuous Environment Benchmark

PreviousDiscrete Environment BenchmarkNextAtari Environment Benchmark

Last updated 5 years ago

Was this helpful?

Continuous Environment Benchmark Result

Env. \ Alg.

A2C (GAE)

A2C (n-step)

PPO

SAC

RoboschoolAnt

787

1396

1843

2915

RoboschoolAtlasForwardWalk

59.87

88.04

172

800

RoboschoolHalfCheetah

712

439

1960

2497

RoboschoolHopper

710

285

2042

2045

RoboschoolInvertedDoublePendulum

996

4410

8076

8085

RoboschoolInvertedPendulum

995

978

986

941

RoboschoolReacher

12.9

10.16

19.51

19.99

RoboschoolWalker2d

280

220

1660

1894

RoboschoolHumanoid

99.31

54.58

2388

2621*

RoboschoolHumanoidFlagrun

73.57

178

2014

2056*

RoboschoolHumanoidFlagrunHarder

-429

253

680

280*

Unity3DBall

33.48

53.46

78.24

98.44

Unity3DBallHard

62.92

71.92

91.41

97.06

Episode score at the end of training attained by SLM Lab implementations on continuous control problems. Reported episode scores are the average over the last 100 checkpoints, and then averaged over 4 Sessions. Results marked with * require 50M-100M frames, so we use the hybrid synchronous/asynchronous version of SAC to parallelize and speed up training time.

Continuous Environment Benchmark Result Plots

Plot Legend

🥇
📈
Upload PR #427
Dropbox data
legend