SAC employs three optimizers for the three losses indicated by the learning rates’ names.
This way they can be tuned independently of each other
Thus, these learning rates specify the rate at which parameters of affected parts of the DNNs change. The exact specification of the loss can be found at ray.rllib.agents.sac.sac_tf_policy.sac_actor_critic_loss. How this loss is composed and what exactly happens there is quite complex and if you don’t want to go through the paper, I would like to suggest this video.
On a high level: Actor and Critic are for actor and critic like you would expect; Entropy loss is an additional SAC loss that stems from SAC’s objective of maximizing not only your rewards but also the entropy of your actions.