I noticed that in the PPO agent initialization it forces the is_action_continuous=False whereas the PPO algorithm and other libraries implementing PPO allow continuous actions. Can this be added to Pearl as well
https://github.com/facebookresearch/Pearl/blob/main/pearl/policy_learners/sequential_decision_making/ppo.py#L99
I noticed that in the PPO agent initialization it forces the
is_action_continuous=Falsewhereas the PPO algorithm and other libraries implementing PPO allow continuous actions. Can this be added to Pearl as wellhttps://github.com/facebookresearch/Pearl/blob/main/pearl/policy_learners/sequential_decision_making/ppo.py#L99