How severe does this issue affect your experience of using Ray?
- High: It blocks me to complete my task.
QMIX this._loss is None
qmix_policy.py
when set num_gpus > 0, the error is following:
2022-09-06 06:54:29,618 ERROR trial_runner.py:980 – Trial QMIX_grouped_twostep_b920e_00000: Error processing event.
ray.exceptions.RayTaskError(ValueError): ray::QMix.train() (pid=73360, ip=172.17.0.11, repr=QMix)
TypeError: ‘NoneType’ object is not callable
The above exception was the direct cause of the following exception:
ray::QMix.train() (pid=73360, ip=172.17.0.11, repr=QMix)
File “/opt/conda/lib/python3.7/site-packages/ray/tune/trainable/trainable.py”, line 347, in train
result = self.step()
File “/opt/conda/lib/python3.7/site-packages/ray/rllib/algorithms/algorithm.py”, line 661, in step
results, train_iter_ctx = self._run_one_training_iteration()
File “/opt/conda/lib/python3.7/site-packages/ray/rllib/algorithms/algorithm.py”, line 2382, in _run_one_training_iteration
recreate=self.config[“recreate_failed_workers”],
File “/opt/conda/lib/python3.7/site-packages/ray/rllib/algorithms/algorithm.py”, line 2190, in try_recover_from_step_attempt
raise error
File “/opt/conda/lib/python3.7/site-packages/ray/rllib/algorithms/algorithm.py”, line 2373, in _run_one_training_iteration
results = self.training_step()
File “/opt/conda/lib/python3.7/site-packages/ray/rllib/algorithms/qmix/qmix.py”, line 274, in training_step
train_results = multi_gpu_train_one_step(self, train_batch)
File “/opt/conda/lib/python3.7/site-packages/ray/rllib/execution/train_ops.py”, line 177, in multi_gpu_train_one_step
permutation[batch_index] * per_device_batch_size, buffer_index=0
File “/opt/conda/lib/python3.7/site-packages/ray/rllib/policy/torch_policy.py”, line 591, in learn_on_loaded_batch
tower_outputs = self._multi_gpu_parallel_grad_calc(device_batches)
File “/opt/conda/lib/python3.7/site-packages/ray/rllib/policy/torch_policy.py”, line 1170, in _multi_gpu_parallel_grad_calc
raise output[0] from output[1]
ValueError: ‘NoneType’ object is not callable
tracebackTraceback (most recent call last):
File “/opt/conda/lib/python3.7/site-packages/ray/rllib/policy/torch_policy.py”, line 1064, in _worker
self._loss(self, model, self.dist_class, sample_batch)
TypeError: ‘NoneType’ object is not callable