You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I suggest expanding the system's recurrent components by introducing various recurrent neural networks (RNNs) like vanilla RNN, GRU, and maybe some lesser-know networks like LMU, and ctRNN. Additionally, I propose compatibility with other RL algorithms beyond PPO, specifically A2C.
Motivation
The motivation is to enhance flexibility, allowing users to choose from a diverse set of recurrent networks and RL algorithms.
Pitch
Introduce different recurrent net options for different RL algorithms such as A2C, providing users with a more comprehensive toolkit for designing and experimenting RL with recurrent components.
Alternatives
Focus on LstmPPO: While effective, this limits exploration and potentially misses out on the strengths of other RNNs.
Develop custom algorithms: This is resource-intensive and may not be as widely applicable as expanding existing options.
Additional context
I have already implemented most of these features in my personal repository and successfully utilized them in my research.
Checklist
I have checked that there is no similar issue in the repo
If I'm requesting a new feature, I have proposed alternatives
The text was updated successfully, but these errors were encountered:
introducing various recurrent neural networks (RNNs) like vanilla RNN, GRU, and maybe some lesser-know networks like LMU, and ctRNN.
I have already implemented most of these features in my personal repository and successfully utilized them in my research.
do you have a benchmark to share?
and are you willing to implement and benchmark those alternatives? (I would start with GRU only at first)
adding more options will add complexity to an already complex algorithm, so we should do that only if it is really beneficial.
🚀 Feature
I suggest expanding the system's recurrent components by introducing various recurrent neural networks (RNNs) like vanilla RNN, GRU, and maybe some lesser-know networks like LMU, and ctRNN. Additionally, I propose compatibility with other RL algorithms beyond PPO, specifically A2C.
Motivation
The motivation is to enhance flexibility, allowing users to choose from a diverse set of recurrent networks and RL algorithms.
Pitch
Introduce different recurrent net options for different RL algorithms such as A2C, providing users with a more comprehensive toolkit for designing and experimenting RL with recurrent components.
Alternatives
Focus on LstmPPO: While effective, this limits exploration and potentially misses out on the strengths of other RNNs.
Develop custom algorithms: This is resource-intensive and may not be as widely applicable as expanding existing options.
Additional context
I have already implemented most of these features in my personal repository and successfully utilized them in my research.
Checklist
The text was updated successfully, but these errors were encountered: