Hey @bug404 . If you have an RLlib Trainer object:
print(trainer.get_policy().model)
If you are using tune.run, try to add this line into ray/rllib/policy/torch_policy.py:~164 (after we assign self.model to the 0th of the multi-GPU towers):
print(self.model)