You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Removed default value "adam" for Tensorforce agent argument optimizer (since default optimizer argument learning_rate removed, see below)
Removed option "minimum" for Tensorforce agent argument memory, use None instead
Changed default value for dqn/double_dqn/dueling_dqn agent argument huber_loss from 0.0 to None
Layers:
Removed default value 0.999 for exponential_normalization layer argument decay
Added new layer batch_normalization (generally should only be used for the agent arguments reward_processing[return_processing] and reward_processing[advantage_processing])
Added exponential/instance_normalization layer argument only_mean with default False
Added exponential/instance_normalization layer argument min_variance with default 1e-4
Optimizers:
Removed default value 1e-3 for optimizer argument learning_rate
Changed default value for optimizer argument gradient_norm_clipping from 1.0 to None (no gradient clipping)
Added new optimizer doublecheck_step and corresponding argument doublecheck_update for optimizer wrapper
Added option to specify agent argument saver as string, which is interpreted as saver[directory] with otherwise default values
Added default value for agent argument saver[frequency] as 10 (save model every 10 updates by default)
Changed default value of agent argument saver[max_checkpoints] from 5 to 10
Summarizer:
Added option to specify agent argument summarizer as string, which is interpreted as summarizer[directory] with otherwise default values
Renamed option of agent argument summarizer from summarizer[labels] to summarizer[summaries] (use of the term "label" due to earlier version, outdated and confusing by now)
Changed interpretation of agent argument summarizer[summaries] = "all" to include only numerical summaries, so all summaries except "graph"
Changed default value of agent argument summarizer[summaries] from ["graph"] to "all"
Changed default value of agent argument summarizer[max_summaries] from 5 to 7 (number of different colors in TensorBoard)
Added option summarizer[filename] to agent argument summarizer
Recorder:
Added option to specify agent argument recorder as string, which is interpreted as recorder[directory] with otherwise default values
run.py script:
Added --checkpoints/--summaries/--recordings command line argument to enable saver/summarizer/recorder agent argument specification separate from core agent configuration
Examples:
Added save_load_agent.py example script to illustrate regular agent saving and loading
Buxfixes
Fixed problem with optimizer argument gradient_norm_clipping not being applied correctly
Fixed problem with exponential_normalization layer not updating moving mean and variance correctly
Fixed problem with recent memory for timestep-based updates sometimes sampling invalid memory indices