Agents and environment implementation #2081
Rakesh-123-cryp
started this conversation in
Ideas
Replies: 1 comment
-
Thanks for reaching out! I see some nice ideas, but I'm missing a bit of context here. Is this part of GSoC? What problem do you try to solve? Do you have specific questions or ideas you want to discuss? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
This below is the wrapper class
This class can be used for any tensor flow or KerasRL agents used and can be also considered for OpenAI Baselines.
Where as for the Environment of the agent we can either let the user integrate the gym environment and project it on the web based visualiser and we can also provide the user with mesa's inbuilt functionality to create environments as such which can also relatively lower the memory usage and possibly speedup the visualisation. When projecting the OpenAI environment onto the visualiser we can also provide them with hyper-parameter tuner to observe the effect in the environment.
We can also create an experience replay buffer library that can hold the episodes of actions, observation, reward and the respective metrics at the end of each episode such as success rate, avg reward, cumulative reward, Short term and Long term Risks, etc.
Beta Was this translation helpful? Give feedback.
All reactions