LLM model loading

Hi everyone,

I have a model that requires 16GB of RAM and 8 CPUs to execute. In my cluster, I have two worker nodes, each with 8GB of RAM and 8 CPUs.

Is it possible to run that task assuming that I have an 16gb of ram combined (4gb+4gb)

Any help or guidance would be greatly appreciated!

THANK YOU!!!