Hello,
I’m new to Ray/RLLib and I have been trying to test scaling basic RL training on my M1 mac air (8 cores). However, whatever I do, it seems the training runs on all cores.
Here’s an example of a config I currently have for PPO:
# using ray.init(num_cpus=4, num_gpus=0)
PPOConfig()
.api_stack(
enable_rl_module_and_learner=True,
enable_env_runner_and_connector_v2=True,
)
.environment(env)
.env_runners(num_env_runners=3,
num_envs_per_env_runner=1,
num_cpus_per_env_runner=1,
num_gpus_per_env_runner=0)
.training(
lr=0.001,
gamma=0.99,
clip_param=0.2,
train_batch_size=200 * 3,
lambda_=0.95,
num_epochs=10,
)
.learners(
num_cpus_per_learner=1,
)
.resources(
num_cpus_for_main_process=1,
)
I was wondering what the problem is, as I have spent hours trying to fix this but with no luck. I’m starting to think it may not be supported on m1, even though I was under the impression it was supposed to be.
Thank you!