I am using the PPO trainer with a custom environment whose observation space is a graph, I am modeling the graph as a GGNN model. Node representations returned from GGNN are given input to the agent as observations. I want the GGNN model to get updated when agent networks are getting updated, how can I integrate GGNN with the agent model?
Thanks for the question Siddharth.
We started to work on a more native support for GNNs in RLlib recently.
At this point though, the best bet may be to just padding your observations to max number of nodes and features. and unpack them in your custom model.
You may find this thread interesting: