How severe does this issue affect your experience of using Ray?
- High: It blocks me to complete my task.
I am attempting to convert a WebLoader into a Ray DataLoader in the open_clip repo. The webloader already have information about how to split between workers, … so I’d like to keep that.
When applying the standard process (wrapping the dataloader into ray.train.torch.prepare_data_loader
, I get this following error:
File ".../ray/train/torch/train_loop_utils.py", line 391, in prepare_data_loader
and not isinstance(data_loader.sampler, DistributedSampler)
AttributeError: 'WebLoader' object has no attribute 'sampler'
So 2 questions here:
- What’s the best way to convert a
wds.WebLoader
to a Ray-compliant data loader? - Less important: How can I carry over the parameters (such as split by worker)?
Thanks!
PS: I’m using the typical TorchTrainer. I see some folks using lower level primitive like this. Is this the recommendation when one wants little refactoring to do?