@@ -178,28 +178,6 @@ def ppo_pong_ae_base():
178178  return  hparams 
179179
180180
181- @registry .register_hparams  
182- def  pong_model_free ():
183-   """TODO(piotrmilos): Document this.""" 
184-   hparams  =  mfrl_base ()
185-   hparams .batch_size  =  2 
186-   hparams .ppo_eval_every_epochs  =  2 
187-   hparams .ppo_epochs_num  =  4 
188-   hparams .add_hparam ("ppo_optimization_epochs" , 3 )
189-   hparams .add_hparam ("ppo_epoch_length" , 30 )
190-   hparams .add_hparam ("ppo_learning_rate" , 8e-05 )
191-   hparams .add_hparam ("ppo_optimizer" , "Adam" )
192-   hparams .add_hparam ("ppo_optimization_batch_size" , 4 )
193-   hparams .add_hparam ("ppo_save_models_every_epochs" , 1000000 )
194-   env  =  gym_env .T2TGymEnv ("PongNoFrameskip-v4" , batch_size = 2 )
195-   env .start_new_epoch (0 )
196-   hparams .add_hparam ("env_fn" , make_real_env_fn (env ))
197-   eval_env  =  gym_env .T2TGymEnv ("PongNoFrameskip-v4" , batch_size = 2 )
198-   eval_env .start_new_epoch (0 )
199-   hparams .add_hparam ("eval_env_fn" , make_real_env_fn (eval_env ))
200-   return  hparams 
201- 
202- 
203181@registry .register_hparams  
204182def  dqn_atari_base ():
205183  # These params are based on agents/dqn/configs/dqn.gin 
@@ -242,7 +220,7 @@ def dqn_original_params():
242220@registry .register_hparams  
243221def  mfrl_original ():
244222  return  tf .contrib .training .HParams (
245-       game = "" ,
223+       game = "pong " ,
246224      base_algo = "ppo" ,
247225      base_algo_params = "ppo_original_params" ,
248226      batch_size = 16 ,
0 commit comments