I am trying to run PPO using a GPU for the trainer.
My setup is the following:
Tensorflow works fine with GPUs. However, when I run the PPO algorithm with “rllib train”, the GPUs are not detected and I get the following error:
RuntimeError: GPUs were assigned to this worker by Ray, but your DL framework (tf) reports GPU acceleration is disabled. This could be due to a bad CUDA- or tf installation.
I tried to remove the part that raised the error, but I noticed that the trainer used only the CPU.
Does anybody know what I could do to fix it?