-
Notifications
You must be signed in to change notification settings - Fork 106
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to export or load multi-agent policies? #49
Comments
Hello, we are very sorry that we have not considered the operation of saving and loading a specific agent model for MARL. At present, saving and loading models in XuanCe are used for the holistic MARL models. We will consider fixing this issue in the future version of XuanCe. Thanks very much for your questions. |
This would be very helpful! I like your overall framework for training and evaluating RL agents and the variety of algorithms, but accessing specific MARL policies for inference is indeed an important feature. |
Thank you for your advice; we will consider supporting this function very soon. |
Additionally, can you also consider adding a feature to train selected agents with a shared policy? e.g. |
It's a good suggestion for more general scenarios, implementing this feature in the current version of XuanCe sounds not easy. We may need to rebuild the policy module of XuanCe to determine whether or not some agents use shared policy. |
Ray RLlib supports both these features, however their variety of algorithms is limited. Anyhow, I would be very thankful if you could provide these features soon. I assume many others would benefit from it. |
In a multi-agent setting, when training e.g.
MAPPO_Agents()
, then callingMAPPO_Agents.save_model(model_name='model.pth')
and finally loading the modelMAPPO_Agents.load_model(path)
, how can I extract specific policies out of the model? e.g. when 3 agents were trained, "agent_1", "agent_2" and "agent_3", I would like to have the prediction only of a specific agent.I didn't find out how to exactly do that and would be very thankful for any help.
The text was updated successfully, but these errors were encountered: