From 0bf97ce1acb9b84ade4ce5d0becc511936354da7 Mon Sep 17 00:00:00 2001 From: akshay-anyscale <122416226+akshay-anyscale@users.noreply.github.com> Date: Tue, 28 May 2024 12:31:51 -0700 Subject: [PATCH] Update README.md Signed-off-by: akshay-anyscale <122416226+akshay-anyscale@users.noreply.github.com> --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index ac8bab19..1865d2b6 100644 --- a/README.md +++ b/README.md @@ -1,7 +1,7 @@ ============================ # Archiving Ray LLM -We had started RayLLM to simplify setting up and deploying LLMs on top of Ray Serve. In the past few months, vLLM has made significant improvements in ease of use. We are archiving the RayLLM project and instead adding some examples to our Ray Serve docs for deploying LLMs with Ray Serve and vLLM. This will reduce another library for the community to learn about and greatly simplify the workflow to serve LLMs at scale. We also recently launched Hosted Anyscale where you can serve LLMs with Ray Serve with some more capabilities out of the box like multi-lora with serve multiplexing, JSON mode function calling and further performance enhancements. +We had started RayLLM to simplify setting up and deploying LLMs on top of Ray Serve. In the past few months, vLLM has made significant improvements in ease of use. We are archiving the RayLLM project and instead adding some examples to our [Ray Serve docs](https://docs.ray.io/en/master/serve/tutorials/vllm-example.html) for deploying LLMs with Ray Serve and vLLM. This will reduce another library for the community to learn about and greatly simplify the workflow to serve LLMs at scale. We also recently launched [Hosted Anyscale](https://www.anyscale.com/) where you can serve LLMs with Ray Serve with some more capabilities out of the box like multi-lora with serve multiplexing, JSON mode function calling and further performance enhancements. ============================