Agent Spec: DDQN+PER on LunarLander
📂The Agent Spec
{
"{spec_name}": {
"agent": [{
"name": str,
"algorithm": {
// Name of an algorithm class in slm_lab/agent/algorithm/
"name": str,
// The probability distribution class used for sampling actions, declared in slm_lab/agent/algorithm/policy_util.py
// - "default": use the default distribution based on the action type (discrete/continuous) of the environment
// - {str}: use a custom distribution
"action_pdtype": str,
// The action policy used, defined in slm_lab/agent/algorithm/policy_util.py
// - "default": directly sample action from the action distribution
// - "random": sample action randomly from the environment's action space
// - "epsilon_greedy": use epsilon-greedy policy (e.g. for DQN family)
// - "boltzmann": use Boltzmann policy (e.g. for DQN family)
"action_policy": str,
// Algorithm-specific options
...
},
"memory": {
// Name of a memory class in slm_lab/agent/memory/
"name": str,
// Memory-specific options
...
},
"net": {
// Name of a network class in slm_lab/agent/net/
"type": str,
// Network-specific options
...
}
}],
"env": [{...}],
...
}
}✍️ Agent Spec for DDQN+PER
🚀 Running DDQN+PER on LunarLander



Last updated
Was this helpful?