RuntimeError: Request failed with status code 500: Traceback (most recent call last):
File “/home/ray/anaconda3/lib/python3.7/site-packages/ray/dashboard/modules/job/job_head.py”, line 269, in upload_package
upload_package_to_gcs(package_uri, await req.read())
File “/home/ray/anaconda3/lib/python3.7/site-packages/aiohttp/web_request.py”, line 655, in read
max_size=self._client_max_size, actual_size=body_size
aiohttp.web_exceptions.HTTPRequestEntityTooLarge: Request Entity Too Large
What is the max size for the request? There is a csv file of 40MB that I copy now which is later used by the python code to retrieve training data. Is it against pattern? I guess the best way is to put the files in a cloud bucket so it doesn’t need to be copied from the process where the job is submitted.
This size should be fine for job submission. Can you share a script that reproduces the problem @Y_C ? I want to see if there was anything special that can cause this.
it is very unlikely that the code size grows into 10s of MB, it is always some hidden files in the working directory folder that often goes unnoticed. So in my experience, it is always a best practice to add ‘excludes’ attribute in the runtime env like the following: