-
Notifications
You must be signed in to change notification settings - Fork 71
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ready for testing 🧪 Multi-policy training support #181
Conversation
I've done a little testing with some of my previous envs and Jumper Hard with older plugin version, and it seemed to work with both multiagent set to LSTM/Attention wrappers work (they show a deprecated warning so possibly accessing them might be different when newer versions of rllib come out), but for exporting we can't use them yet since the state data wouldn't be fed in. One thing I found that doesn't work well is enabling some exploration options with PPO, one that worked was RE3 with Tensorflow rather than Torch set. Curiosity needs discrete or multidiscrete actions, but didn't seem to work when I switched the env to discrete actions. I think it might be related to the tuple action space, it might not be supported in some of the exploration codes. Warning Edit: With the current script, the exported onnx from Rllib doesn't output just action means like our SB3 setup, so the output size is doubled and exported onnx with more than one action won't work correctly. Edit2: I've just updated the plugin to handle the case above. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM pending review of other PRs
|
||
if __name__ == "__main__": | ||
parser = argparse.ArgumentParser(allow_abbrev=False) | ||
parser.add_argument("--config_file", default="rllib_config.yaml", type=str, help="The yaml config file") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think this should be examples/rllib_config.ymal
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I usually call the example from within the examples folder, so the default was based on my usage. If calling from GDRL repository directly then it should be changed.
If someone installs GDRL using pip install
and then just downloads the example file and config file, they might not have the entire repository, but I'm not sure how common this is.
I leave this up to you, I can definitely change the default.
- Adds supports for exporting envs with multidiscrete actions with sb3 - Multiple obs spaces onnx export support (for sb3) still needs to be worked on in the future
Also removes the previously removed init variables from `tune.register_env()`
Updates rllib doc to include the new process.
Adds support for training multiple policies with Rllib.
Plugin PR: edbeeching/godot_rl_agents_plugin#40
Example env PR: edbeeching/godot_rl_agents_examples#30
TODO: