aiohttp.web_exceptions.HTTPRequestEntityTooLarge: Request Entity Too Large

I am reading data from csv file around 200 MB size and 10lakh records and putting in pandas dataframe and creating pdf’s from dataframe, I have cluster of 4 nodes while submitting ray job this error arise,

aiohttp.web_exceptions.HTTPRequestEntityTooLarge: Request Entity Too Large

The error aiohttp.web_exceptions.HTTPRequestEntityTooLarge: Request Entity Too Large occurs because Ray enforces a 100 MiB limit on the working directory or files uploaded with a job submission. If your working directory or files (including the CSV or generated PDFs) exceed this limit, the upload fails with this error. The recommended solution is to store large files (like your CSV or PDFs) in cloud object storage (e.g., S3) and access them from within your Ray job, rather than uploading them directly with the job submission. You can also use the excludes attribute in runtime_env to prevent large files from being uploaded with the job submission, or use a remote URI for the working directory if needed. See examples and discussion in the sources below.

Would you like more detail on how to restructure your workflow or configure Ray to avoid this error?

Sources:

Hint: Mention @RunLLM in the post for followups.

I am using in premises storage on rhel8 server, my server does not have internet ,so i can’t use remote directory, can you please suggest how can i handle this issue for large files on local storage @RunLLM