fix persist a single training environment for PPO

This commit is contained in:
sonnhfit
2022-08-19 01:49:11 +07:00
committed by robcaulk
parent f95602f6bd
commit 4baa36bdcf
3 changed files with 51 additions and 25 deletions

View File

@@ -79,13 +79,9 @@
"random_state": 1,
"shuffle": false
},
"model_training_parameters": {
"model_training_parameters": {
"learning_rate": 0.00025,
"gamma": 0.9,
"target_update_interval": 5000,
"buffer_size": 50000,
"exploration_initial_eps":1,
"exploration_final_eps": 0.1,
"verbose": 1
},
"rl_config": {
@@ -103,4 +99,4 @@
"internals": {
"process_throttle_secs": 5
}
}
}