Hello,
I encounter this error while passing the config to PPO:
(MultiAgentEnvRunner pid=28036) File “D:\Miniconda\envs\condaPars\Lib\site-packages\ray\rllib\core\rl_module\rl_module.py”, line 113, in build
(MultiAgentEnvRunner pid=28036) module = self.module_class(module_config)
(MultiAgentEnvRunner pid=28036) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(MultiAgentEnvRunner pid=28036) deprecation_warning(
(MultiAgentEnvRunner pid=28036) File “D:\Miniconda\envs\condaPars\Lib\site-packages\ray\rllib\utils\deprecation.py”, line 48, in deprecation_warning
(MultiAgentEnvRunner pid=28036) raise ValueError(msg)
(MultiAgentEnvRunner pid=28036) ValueError: RLModule(config=[RLModuleConfig])
has been deprecated. Use RLModule(observation_space=.., action_space=.., inference_only=.., learner_only=.., model_config=..)
instead.
# Create PPO configuration with default policies
self.config = {
"env": "network_env",
"framework": "torch",
"num_gpus": 0,
"num_workers": 2,
"train_batch_size": 4000,
"sgd_minibatch_size": 128,
"lr": 0.0003,
"gamma": 0.99,
"lambda": 0.95,
"use_gae": True,
"disable_env_checking": True,
"_disable_preprocessor_api": True,
"_disable_action_flattening": True,
"experimental": {
"_disable_rl_module": True, # این خط را اضافه کنید
"_validate_config": False,
"enable_rl_module_and_learner": False,
"enable_env_runner_and_connector_v2": False
},
"clip_param": 0.2,
"vf_loss_coeff": 1.0,
"entropy_coeff": 0.0,
"num_sgd_iter": 30,
"rollout_fragment_length": 200,
"batch_mode": "truncate_episodes",
"model": {
"fcnet_hiddens": [256, 256],
"fcnet_activation": "tanh",
},
"multiagent": {
"policies": {
"orchestrator_policy": (None, env.orchestrator_observation_space, env.orchestrator_action_space, {}),
"controller_policy": (None, env.controller_observation_space, env.controller_action_space, {})
},
"policy_mapping_fn": lambda agent_id, *args, **kwargs: (
"orchestrator_policy" if agent_id.startswith("orchestrator")
else "controller_policy"
)
},
"observation_filter": "MeanStdFilter",
"create_env_on_driver": False,
"log_level": "INFO",
# Disable experimental validation and new APIs
"experimental": {
"_validate_config": False,
"enable_rl_module_and_learner": False,
"enable_env_runner_and_connector_v2": False
}
}
# Create trainer
# Create trainer
logger.info("Creating PPO trainer...")
try:
# Initialize trainer
self.trainer = PPO(config=self.config)
logger.info("PPO trainer created successfully")
except Exception as e:
logger.error(f"Error creating PPO trainer: {str(e)}")
raise e
How can I address this problem???