-
-
Notifications
You must be signed in to change notification settings - Fork 112
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Does it allow defining an environment that has continuous action space? And how? #1078
Comments
Yes, but how to model the action space mainly depends on your problem. For example, you can use negative log-normal distribution. Or you can split the action space into discrete bins then randomly sample from it. |
Thank you for your reply, @findmyway !
But I wonder how should I split the action space? Since the action space is not explicitly given, but is implicitly determined by a set of constraints, how do I obtain and split this action space? By the way, could you recommend some tutorials that are suitable for learning reinforcement learning and this package at the same time? Thank you very much! |
It's hard to answer without further information here.
I'd love to, but since I haven't done any RL related work recently, I'd leave it for others to answer it. |
OK. Thank you @findmyway ! 🤝 |
First of all, I would like to say thank you to all of the contributors of this useful package!
I am a learner of both RL and this package. I wonder if RL or this package can deal with problems that have a continuous action space or a mixed integer action space. Specifically, suppose we have a decision making problem that at each step we are to make a decision (an action)$a$ according to current state $s$ and an observation of the random noise $\omega$ , i.e., the action space $\mathcal{A}$ is characterized by some constraints such as $\mathcal{A} = { a| f(a, s, \omega) \le 0 }$ (BTW, why can't github display
{}
?). Can RL deal with this kind of problems? And how can I write such a environment usingRenforcementLearning.jl
?Looking forward to your reply at your convenience! Thanks!
The text was updated successfully, but these errors were encountered: