You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I get no evaluations at all. The problem is two-fold:
-EE is being compared to history["episodes"] which for the ppo agent now measures episodes as determined by the -E argument times number of rollouts. In this case, I run 10000 'episodes' but for every episode history["episodes"] is incremented by 200. This means to get the desired behavior, I have to set -EE 2000. This is very unexpected behavior and inconsistent with the other agents.
Even when setting -EE 2000, I still get no evaluations. This seems to be caused by the check that is performed to determine whether we should evaluate:
After changing the way episodes are tracked for the ppo agent (#58), the
eval_every
argument does not longer work as expected. If I run for example:I get no evaluations at all. The problem is two-fold:
-EE
is being compared to history["episodes"] which for the ppo agent now measures episodes as determined by the-E
argument times number of rollouts. In this case, I run 10000 'episodes' but for every episodehistory["episodes"]
is incremented by 200. This means to get the desired behavior, I have to set-EE 2000
. This is very unexpected behavior and inconsistent with the other agents.Even when setting
-EE 2000
, I still get no evaluations. This seems to be caused by the check that is performed to determine whether we should evaluate:This does not work anymore because
history["episode"]
does no longer increase in increments of 1 between these checks. So I will have the following:and
eval_next
will always beFalse
. I now changed the check to the following as a quick fix:This works for the experiments, I want to run. However, I am not sure if this is a good solution in general and we should discuss what to do here.
The text was updated successfully, but these errors were encountered: