About the LLMs/Generative AI/Aviary category
|
|
0
|
520
|
April 20, 2023
|
GPT-J6B Sample Code
|
|
0
|
30
|
April 10, 2024
|
RayTaskError(OutOfMemoryError) when using a LLM
|
|
0
|
28
|
April 2, 2024
|
Installing TensorRT LLM on Ray Docker Image as Custom Docker
|
|
2
|
342
|
March 7, 2024
|
Best practices to run multiple models in multiple GPUs in RayLLM
|
|
0
|
343
|
February 8, 2024
|
Question about the model yaml config `accelerator_type_a100`
|
|
1
|
353
|
February 2, 2024
|
Overriding resources per worker in ray-llm
|
|
7
|
245
|
January 30, 2024
|
How to assign actors to specific machines?
|
|
2
|
147
|
January 8, 2024
|
How to deploy LLM models that can handle high concurrency based on the Ray serve framework
|
|
1
|
619
|
January 8, 2024
|
Can RayLLM be deployed on Azure Cloud Services?
|
|
1
|
131
|
December 15, 2023
|
Download an opensource LLM model in Raycluster yaml file?
|
|
2
|
164
|
December 14, 2023
|
OOM when I decoupled ray from GPTj finetune script
|
|
0
|
179
|
November 17, 2023
|
Does ray-llm support only CPU?
|
|
0
|
288
|
October 25, 2023
|
Memory Requirements for distributing LLM
|
|
3
|
440
|
August 31, 2023
|
Turbocharge LangChain: guide to 20x faster embedding
|
|
14
|
2962
|
July 27, 2023
|
Use deepspeed in aviary to deploy falcon 40B / Llama 30B Fails
|
|
3
|
808
|
July 23, 2023
|
How to deploy LLaMA 2 7B model with Aviary
|
|
0
|
1632
|
July 20, 2023
|
HuggingFacePredictor Multi-GPU
|
|
3
|
629
|
July 12, 2023
|
How to deploy Aviary Frontend along with Backend in the same Ray cluster?
|
|
1
|
560
|
June 9, 2023
|
Request for comments: Adding LangChain support to Aviary
|
|
0
|
440
|
June 2, 2023
|
Announcing Ray Aviary
|
|
1
|
657
|
May 31, 2023
|