Are these two ways of training an agent equivalent in terms of number of steps and exploration strategies?
analysis = ray.tune.run(dqn.DQNTrainer, config=self.config, local_dir=self.save_dir, stop={"training_iteration": 100},
checkpoint_at_end=True)
for n in range(100):
result = agent.train()