Skip to content

Commit

Permalink
Fix vllm typo
Browse files Browse the repository at this point in the history
  • Loading branch information
damccorm authored Nov 15, 2024
1 parent 7650a6d commit cc57bde
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion examples/notebooks/beam-ml/run_inference_vllm.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -66,7 +66,7 @@
{
"cell_type": "markdown",
"source": [
"[vLLM](https://github.com/vllm-project/vllm) is a fast and user-frienly library for LLM inference and serving. vLLM optimizes LLM inference with mechanisms like PagedAttention for memory management and continuous batching for increasing throughput. For popular models, vLLM has been shown to increase throughput by a multiple of 2 to 4. With Apache Beam, you can serve models with vLLM and scale that serving with just a few lines of code.\n",
"[vLLM](https://github.com/vllm-project/vllm) is a fast and user-friendly library for LLM inference and serving. vLLM optimizes LLM inference with mechanisms like PagedAttention for memory management and continuous batching for increasing throughput. For popular models, vLLM has been shown to increase throughput by a multiple of 2 to 4. With Apache Beam, you can serve models with vLLM and scale that serving with just a few lines of code.\n",
"\n",
"This notebook demonstrates how to run machine learning inference by using vLLM and GPUs in three ways:\n",
"\n",
Expand Down

0 comments on commit cc57bde

Please sign in to comment.