revision
This commit is contained in:
parent
d43fa600a0
commit
bd40ffc53c
|
@ -146,7 +146,7 @@ def eval_policy(
|
||||||
thread.join()
|
thread.join()
|
||||||
|
|
||||||
info = {
|
info = {
|
||||||
"micro": [
|
"per_episode": [
|
||||||
{
|
{
|
||||||
"episode_ix": i,
|
"episode_ix": i,
|
||||||
"sum_reward": sum_reward,
|
"sum_reward": sum_reward,
|
||||||
|
@ -164,7 +164,7 @@ def eval_policy(
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
],
|
],
|
||||||
"macro": {
|
"aggregated": {
|
||||||
"avg_sum_reward": np.nanmean(sum_rewards[:num_episodes]),
|
"avg_sum_reward": np.nanmean(sum_rewards[:num_episodes]),
|
||||||
"avg_max_reward": np.nanmean(max_rewards[:num_episodes]),
|
"avg_max_reward": np.nanmean(max_rewards[:num_episodes]),
|
||||||
"pc_success": np.nanmean(successes[:num_episodes]) * 100,
|
"pc_success": np.nanmean(successes[:num_episodes]) * 100,
|
||||||
|
@ -218,7 +218,7 @@ def eval(cfg: dict, out_dir=None, stats_path=None):
|
||||||
max_steps=cfg.env.episode_length,
|
max_steps=cfg.env.episode_length,
|
||||||
num_episodes=cfg.eval_episodes,
|
num_episodes=cfg.eval_episodes,
|
||||||
)
|
)
|
||||||
print(info["macro"])
|
print(info["aggregated"])
|
||||||
|
|
||||||
# Save info
|
# Save info
|
||||||
with open(Path(out_dir) / "eval_info.json", "w") as f:
|
with open(Path(out_dir) / "eval_info.json", "w") as f:
|
||||||
|
|
|
@ -183,7 +183,7 @@ def train(cfg: dict, out_dir=None, job_name=None):
|
||||||
video_dir=Path(out_dir) / "eval",
|
video_dir=Path(out_dir) / "eval",
|
||||||
save_video=True,
|
save_video=True,
|
||||||
)
|
)
|
||||||
log_eval_info(logger, eval_info["macro"], step, cfg, offline_buffer, is_offline)
|
log_eval_info(logger, eval_info["aggregated"], step, cfg, offline_buffer, is_offline)
|
||||||
if cfg.wandb.enable:
|
if cfg.wandb.enable:
|
||||||
logger.log_video(first_video, step, mode="eval")
|
logger.log_video(first_video, step, mode="eval")
|
||||||
logging.info("Resume training")
|
logging.info("Resume training")
|
||||||
|
|
Loading…
Reference in New Issue