How severe does this issue affect your experience of using Ray?
- Medium: It contributes to significant difficulty to complete my task, but I can work around it.
Hi All thanks for reading,
After launching a ray cluster in AWS via ray up on latest version, I am attempting to run the vLLM example.
https://docs.ray.io/en/latest/serve/tutorials/vllm-example.html
Serve will start and run the LLM fine. When the query.py is run against it however it has error output, the key piece being below:
(ServeReplica:default:VLLMDeployment pid=286, ip=) AttributeError: ‘str’ object has no attribute ‘name’
I was able to get it to work but only after modification in vLLM code https://github.com/vllm-project/vllm/blob/main/vllm/entrypoints/openai/serving_engine.py:
def _is_model_supported(self, model_name):
return any(model.name == model_name for model in self.base_model_paths)
def _is_model_supported(self, model_name):
return True
And /vllm/blob/main/vllm/entrypoints/openai/serving_chat.py:
line 265:
– model_name = self.base_model_paths[0].name
++ model_name = self.base_model_paths[0]
line 599:
– model_name = self.base_model_paths[0].name
++ model_name = self.base_model_paths[0]
After this the example is working fine and serving queries without errors or issues.
This seems to be a bug in Ray integration with the latest version of vLLM however I am very new with Ray and this is my first effort using it so I am wondering if anyone else is facing the same issue or can validate they are running the latest vLLM example from the official Ray docs without issue on latest version of Ray?